---
id: "concept-context-sprawl"
type: "concept"
source_timestamps: ["00:05:12", "00:15:50"]
tags: ["prompt-engineering", "model-performance"]
related: ["concept-token-burning", "concept-gather-vs-focus"]
definition: "The negative compounding effect of maintaining long, multi-turn LLM conversations, which exponentially increases token costs and dilutes the model's attention and reasoning capabilities."
sources: ["s45-claude-limit-chatgpt-habit"]
sourceVaultSlug: "s45-claude-limit-chatgpt-habit"
originDay: 45
---
# Context Sprawl

## Definition
Context sprawl is the negative compounding effect of letting a single LLM chat session run for too many turns (20, 30, 40+). Costs grow exponentially **and** model reasoning degrades.

## Why It Happens
Users treat chatbots like continuous human conversations, but LLMs are **stateless** — see [[prereq-stateless-architecture]]. With every new prompt the chat client re-submits the *entire* prior history. A simple 'follow-up question' on turn 30 is actually paying to re-process tens of thousands of tokens of prior dialogue, including:
- Every previous mistake and dead-end
- All system prompts
- All ingested documents (especially painful if they weren't converted via [[concept-markdown-conversion]])
- Every tool/plugin schema (see [[concept-silent-tax]])

## Why It Hurts Reasoning, Not Just Costs
The speaker emphasizes that frontier models are generally **not RLHF-trained** to handle massive 40-turn meandering sprawls. As context fills up:
- The ratio of original-critical-instruction to accumulated-noise gets compressed
- Attention is diluted by past dead-ends and irrelevant tangents
- This is consistent with the 'lost in the middle' research (TMLR 2024) showing retrieval accuracy can drop ~50% in mid-context

This is the empirical engine behind the contrarian claim that [[contrarian-models-plateauing]] is an illusion produced by sprawl, not by the models themselves.

## The Counter-Practice
Aggressively summarize, then start fresh. This is operationalized by [[concept-gather-vs-focus]] and [[framework-clean-conversation]], and turned into a habit via [[action-start-fresh-chats]] (≤10–15 turn rule).

## Related Costs
Context sprawl is the second of the three anti-patterns of [[concept-token-burning]] and is one of the items audited by [[framework-stupid-button-audit]].
