Context Window: Why AI "Forgets"

Day 11 · Week 3 · Context — The Secret Sauce

Slack Message — copy & paste

🤖 Tip #11 — AI gets confused late in long sessions because it literally can't see older messages anymore. Start fresh for new tasks.

#11 All Tools

How Context Windows Work

  • Context window sizes vary by model
    • Claude Opus: ~1M tokens (~750K words)
    • Claude Sonnet: ~200K tokens (~150K words)
    • GPT-5.x models: ~128K–200K tokens depending on variant
    • Gemini 2.5 Pro: ~1M tokens
  • When your conversation fills the window, older messages get compressed or droppedthe AI literally can't see them anymore.
  • Symptoms of context overflow
    • AI repeats work it already did
    • AI 'forgets' decisions you agreed on
    • AI contradicts its earlier responses
    • Code quality degrades as the session goes on
  • The mental modelcontext window = short-term memory. CLAUDE.md = long-term memory. Write down what matters.
1M Tokens 200K Tokens Fresh Sessions
#11 Context — The Secret Sauce

How to Manage It

  1. 1 Prevention strategies
    • Start fresh for new tasks — don't chain unrelated work in one session
    • Use subagents for expensive operations — they get their own context
    • Put conventions in CLAUDE.md — re-loaded automatically, never forgotten
    • Compact your context when sessions get long — it summarizes and frees space (but loses nuance)
  2. 2 Watch your token usagestart a new session when you switch to a different task
  3. 3 Move important conventions and decisions into CLAUDE.md so they survive across sessions
  4. 4 Compact context when a session gets long but you're not ready to start freshit summarizes and frees space
1M Tokens 200K Tokens Fresh Sessions
Your screenshot here Optional — add a screenshot from your own workflow
KAI by Dragan Filipovic