← All articles Principles

LLM, not AI: why the terminology matters for how you work

LLM, not AI: why the terminology matters for how you work

The word you use changes how you think

Call it "AI" and you expect it to figure things out. You give it vague instructions and hope it reasons through to the right answer. You assume it catches things you missed. When it fails, you blame the prompt.

Call it an LLM — a large language model — and you understand what is actually happening. You know what to put in. You know why it sometimes gives confident-sounding wrong answers. You stop treating it as a thinking partner and start treating it as an extraordinarily capable tool that does exactly what it is designed to do: predict the most probable continuation of what you gave it.

That shift in understanding changes everything about how you work with it.


What an LLM actually is

A large language model is a text predictor. Given a sequence of tokens — words, code, data — it calculates the statistically most probable next token, then the next, then the next, until the output is complete.

That is the entire mechanism. It is not reasoning in the way you reason. It is not searching a database of facts. It is not maintaining a world model that it updates as it learns. It is completing a pattern from what already exists in the context window — which is everything you gave it, plus everything it generated so far.

The context window is the workspace. What is in it shapes every prediction that follows. An empty context window means the model predicts from general training data patterns alone. A context window filled with your codebase, your architecture decisions, your constraints, and your specific question means the model predicts from all of that.

This is why "context first, prompt second" is not a tip — it is the fundamental operating principle. The model does not go and find information. It predicts from what is already there. You are responsible for making sure the right information is there.


What "AI" implies that LLMs do not deliver

When people say AI, they generally mean something like: a system with general intelligence that can reason about problems, update its beliefs with new information, recognize when it is wrong, and exercise judgment that transfers across domains.

Large language models do not do this.

They do not reason — they pattern-match at extraordinary scale and speed. They do not update — each context window is a fresh start with no memory of previous conversations. They do not recognize when they are wrong — they generate with the same syntactic confidence regardless of whether the content is accurate. Their judgment does not transfer in the way experience transfers; it is statistical correlation across training data, not understanding.

This matters operationally. A system that "reasons" would know when your question is ambiguous and ask for clarification. An LLM generates a plausible-sounding answer to what it inferred your question might mean. A reasoning system would flag uncertainty. An LLM states uncertain things with the same fluency as certain things.

The single most common LLM failure mode — generating a confident, fluent, completely wrong answer — is only surprising if you expected a reasoning agent. It is entirely expected behavior from a text predictor.


Why it matters for accountability

The terminology has a direct consequence for accountability.

If you think you are working with "AI," it is easy to treat its output as something produced by an external intelligence — something to be delivered, evaluated, and passed on. The AI did the analysis. The AI built the feature. The AI wrote the report.

If you know you are working with an LLM, you know that every output is a pattern completion based on what you put in. The quality of the output reflects the quality of the context you provided, the specificity of what you asked, and your judgment in evaluating the result. There is no external intelligence to credit or blame. There is a tool, and there is you.

That is not a subtle distinction. It is the difference between "the AI got it wrong" and "I accepted output I should have reviewed more carefully." One of those is a sentence that prevents learning. The other is not.


What changes when you use the right word

In practice, calling it an LLM — and understanding what that means — produces three immediate changes:

You load context deliberately. Instead of writing a detailed prompt and hoping the model figures out the rest, you ask: what does this model need in order to predict the right output? You load the relevant files, the relevant history, the relevant constraints. Then you ask.

You evaluate output as a prediction, not a decision. The model generated the most probable continuation of what it was given. Is that probable continuation actually correct? That is your question to answer — and you can only answer it if you know enough about the subject matter to recognize a wrong answer.

You stop looking for a better prompt to fix things that context would fix. Most "prompt engineering" problems are context problems. The model does not have what it needs to produce a useful output. Adding more information to the context window solves them. A cleverer prompt does not.


The practical boundary

None of this means LLMs are not remarkable. They are. The pattern completion happens at a scale and quality that was not possible five years ago, and the applications are genuine.

A developer who understands what an LLM is — and uses it accordingly — can work at a multiplier that was not previously achievable. A developer who treats it as an autonomous AI agent, and ships its output without verification, gets a different result.

The tool is the same. The understanding of what it is determines the outcome.