The context window is not a dump truck
Modern LLMs can process enormous amounts of text. Context windows that once maxed out at 4,000 tokens now routinely accommodate hundreds of thousands, and some models are approaching the million-token range. This has led to a tempting assumption: If the model can hold more, just put more in. Pipe in all the documents, all the chat history, all the tool outputs, and let the model sort it out.
This approach fails in several distinct ways. The most straightforward is cost: Large contexts are expensive to process and add latency to every request. But the subtler failures are more dangerous.
Research has consistently shown that model accuracy degrades when context is bloated with irrelevant information. The model doesn't just ignore noise; it gets confused by it. This is sometimes called the “lost in the middle” problem, where information buried deep in a long context is less likely to be retrieved accurately than information near the beginning or end.
There are three failure patterns worth understanding in detail:
- Context poisoning: In a multi-step agentic workflow, the model's outputs from one step become inputs to the next. If the model makes an incorrect assumption early on, that error propagates forward. Each subsequent step reasons on top of a flawed premise. By step six, the system may be confidently executing a task entirely disconnected from what was actually requested.
- Context distraction: More information isn't always better. When agents have access to too many tools with overlapping capabilities, or when documents are loaded without filtering for relevance, the model struggles to identify the signal within the noise. Accuracy drops not because the right information is absent, but because it's buried.
- Context clash: This occurs when the context contains contradictory information, such as a policy document that's been updated since the version in the knowledge base, or a tool output that conflicts with a retrieved fact. Models don't handle contradiction gracefully. They don't flag it; they pick a side and proceed, often unpredictably.

