Definition
Long context refers to AI models that can handle extended input lengths, from hundreds of thousands to millions of tokens.
Context Window Evolution: - GPT-3 (2020): 4K tokens - GPT-4 (2023): 128K tokens - Claude 3 (2024): 200K tokens - Gemini 1.5 (2024): 1M+ tokens
Enabling Technologies: - Efficient attention mechanisms - Positional encoding improvements - Memory optimization - Sparse attention patterns
Use Cases: - Entire codebase analysis - Book-length document Q&A - Long meeting transcripts - Multi-document synthesis
Challenges: - Compute cost scales with length - "Lost in the middle" phenomenon - Retrieval may still outperform
Examples
Uploading an entire novel to Claude and asking it to identify all plot inconsistencies.
Related Terms
Want more AI knowledge?
Get bite-sized AI concepts delivered to your inbox.
Free intelligence briefs. No spam, unsubscribe anytime.