Back to Glossary
concepts

Long Context

AI models capable of processing very large amounts of text in a single prompt.

Share:

Definition

Long context refers to AI models that can handle extended input lengths, from hundreds of thousands to millions of tokens.

Context Window Evolution: - GPT-3 (2020): 4K tokens - GPT-4 (2023): 128K tokens - Claude 3 (2024): 200K tokens - Gemini 1.5 (2024): 1M+ tokens

Enabling Technologies: - Efficient attention mechanisms - Positional encoding improvements - Memory optimization - Sparse attention patterns

Use Cases: - Entire codebase analysis - Book-length document Q&A - Long meeting transcripts - Multi-document synthesis

Challenges: - Compute cost scales with length - "Lost in the middle" phenomenon - Retrieval may still outperform

Examples

Uploading an entire novel to Claude and asking it to identify all plot inconsistencies.

Want more AI knowledge?

Get bite-sized AI concepts delivered to your inbox.

Free intelligence briefs. No spam, unsubscribe anytime.

Discussion