-
Markmap vs Mermaid for Spec-Driven Development (SDD)
Markmap vs Mermaid for Spec-Driven Development You will consider sooner or later. When You have to go „full on AI” documentation stops being a nice-to-have and becomes a must be part of the product. We need to convey a lot of information, with hopefully as little text as possible so we can easily read it and digest the information. Two popular options for Markdown documentation features are Markmap and Mermaid. They solve different problems, and the best choice depends on whether you want fast idea mapping or more structured diagrams for flows. When to use Markmap Markmap turns Markdown headings and bullet points into an interactive mind map. Useful for…
-
Need to know old boy. Principle of Minimum Access for LLMs
Principle of Minimum Access for LLMs can be described by Timothy Dalton as James Bond in „The Living daylights”. Bond says the phrase “Sorry old boy, section 26, paragraph 5, need to know.” to a fellow agent and drives off escorting a VIP – very important target. Behind the scene is a practical idea that fits modern AI systems very well: an LLM or agent (as in the movie) should only be given the minimum access it needs to do its job. Not more, not less. Bare minimum. Of course Mythos probably could jailbreak anyway but still… controll is the best form of trust ? This is not just a…
-
Structure for instructions, agents and skills
This all is the provide a nice clean idea on how to store your files so You can make Your AI assistant / LLM network understand what and how to do. Acording to Your more or less strict rules. This should help You to achieve more repeatable results as expected. The problem : generic answer Do not confuse with generic functions, those rock ! Out-of-the-box, any LLM (here copilot) creates generic code. You could call it 'vanilla’ flavour. It doesn’t know your conventions, library preferences patterns / anti patterns. This results in something that might work but is hard to maintain, totally different then the rest of the lot and…
-
Protect your work with poison ?
Protect your work with poison ? Is there no other way around. Robots.txt ? Nobody cares. Copyrights ? LOL. Fair use policy ? As long as i don`t get caught. Protect your work with poison is the oldest trick in the book, especially by plants. Most of them have to be cooked, to be eaten and digested with benefit for us. Why not doing it with our work ?I am thinking about healthy amount of protecting our work. Does it mean anyone making a rembrandt style photo should pay the author of that style ? For 70 years ? How much and how long ? Why money at all ?…
-
Yet another basic AI glossary part 1
AI & Machine Learning Glossary for Beginners Yet another basic AI glossary part 1. This is base what i need to learn better and understand all that „AI” and „LLMs”. Feel free to go through all of it and dive deeper on those subjects. Defining here ai concepts, ideas, math functions, slang and anything that might be helpful in better understanding „the whole lot”. 1. Logit A logit is the raw output we get from a model before any functions are applied. Before the softmax functions. During classification logit shows the confidence of the model about everypossible output. 2. Logit Definition (Mathematical View) In math logits are real numbers output…
-
How to reduce github copilot`s premium requests usage and maximize efficiency
How to reduce github copilots premium requests usage and maximize efficiency ? Make a plan, a kaizen plan at best. Instruct precisely, cover edge cases, allow all tools to execute and pray the LLM will understand You. Want to share my simple methodology that not only can save money but also ease in and smoothen out the workflow. RTFM ! As always You could benefit from RTFM ! Reading the foqing / friendly / flopsy manual. I know You never read it cause real man don`t do it ( how about real woman ? ) ? God knows if gamers would not have to go through the tutorial, they would…
-
Statics VS AI code analysis ~13 tools
Statics VS AI code analysis works best using the pros from both words. Go hybrid ! Static tools understand the syntext, hardcoded parameters and are very strict. On the other hand AI understands context, can figure out business logic, adapt the codebase. Logic flaws or performance bottlenecks rule-based scanners might miss, AI will put more effort into that. Static analysis limits Static tools scan for syntax errors, style violations, and basic security patterns using fixed rules. Always consistenst, very fast but might generate false positives, ignore business logic, and require manual rule overrides. How often did You use @typescript-error 🙂 Do You code for the linter to pass, logic to…
-
Code mode for mcp servers and llms
Code mode for mcp servers is about LLM writing and calling the code to use a proper MCP instead of calling it directly with the whole context. It makes the call a lot smaller, no overhead is passed, just the basics that are required to call the proper MCP method. Just as You do in code, method or a function, proper parameters, everything validated and… BAM ! We are returning a context that LLM uses in further stages. Anthropics wrote… This is really nice looking but only for a huge models with a 1kk tokens of a context. We need to remember that this is not possible on any kind…
-
Tokenization and embedding of song lyrics „We will, we will…”
Tokenization and embedding of song lyrics „We will, we will…” i know you know how it ends. but have You wondered what would an LLM say ? Let us find out. I want to ask the Clause Sonnet 4.5 about the embeddings, tokenizations and probability of figuring out the lyrics for „We will, we will…” prompt 🙂 Can you show me the tokenization, embeddings, emtadata and probabilities for „We will, we will …” Tokenization Any kind of machine learning uses numbers in the back stage. The text is split into tokens, each mapped to an ID ( example values ) : Token ID Token Text Type Position Length 1234 We…
-
What is a token in AI ? Prompt examples
What is a token in AI ? – A piece of text, usually a word that we send to the LLM. The same goes for when we get a response ( usually number of words = number of tokens, roughly). How to use it ? Since we pay for it , rather cautiously. We want to send the minimum and get the maximum out of every request. Pretty much the basics of economics. Below are couple of prompts i used to analyze my tokens usage. Something are obsiously hallucinations but on the other hand we get a prety decent breakdown of all the data i did send for that coding…





















