Summary
Modern agent systems rely on large language models as flexible decision engines. To understand how those systems work, you do not need every detail of model training. You need the small set of ideas that explain why LLMs are good at instructions, planning, tool selection, and synthesis.Why It Matters
LLMs are often treated like magic at the product layer. That leads to poor design decisions.- some teams assume the model remembers more than it does
- some assume it reasons symbolically when it is really pattern-driven
- some treat context as free even though it is the main operating budget
Mental Model
Four ideas are enough for most agent work.next-token prediction: the model is trained to continue sequences, not to execute a symbolic proof system.tokenization and embeddings: the model operates on tokenized inputs mapped into vector space rather than on raw human-readable text.transformer attention: the model decides what parts of the input matter when generating the next step.pretraining plus adaptation: broad capabilities come from large-scale pretraining, while task performance depends heavily on prompts, context, tools, and any later alignment or tuning.
Architecture Diagram
Tool Landscape
The model properties that matter most for agents are practical ones:- instruction following
- long-context handling
- structured output reliability
- tool-call formatting
- summarization and synthesis quality
Tradeoffs
- Larger models may reason better, but they raise cost and latency.
- Longer context windows reduce some retrieval pressure, but they do not remove the need for context engineering.
- Stronger instruction following helps tool use, but it does not guarantee factual correctness.
- Pretrained priors are broad, but they are not the same thing as current, source-backed knowledge.
- treat the model as a flexible planner and language engine
- use external tools for exact data, computation, and system action
- design around context limits instead of pretending they do not matter
Citations
- Source input: Chapter 3 Fundamentals of Large Language Models
- Source input: Hello-Agents reference boundary
Reading Extensions
Update Log
- 2026-04-21: Initial repo-native draft based on imported reference material and lab rewrite rules.