| Gemini 1.5 Pro | Google | 2M tokens | 2024 | ~1.5M words | 8,192 out | Largest publicly available context window. |
| GPT-4.1 | OpenAI | 1M tokens | 2025 | ~750K words | 32,768 out | Long-context variant for big-document workflows. |
| Gemini 1.5 Flash | Google | 1M tokens | 2024 | ~750K words | 8,192 out | Fast, efficient long-context model. |
| Gemini 2.0 Flash | Google | 1M tokens | 2024 | ~750K words | 8,192 out | Improved Flash with native tool use and multimodal output. |
| Grok 3 | xAI | 1M tokens | 2025 | ~750K words | Not published | Long-context successor with stronger reasoning. |
| Codestral | Mistral | 256K tokens | 2024 | ~192K words | Not published | Specialized code model, larger context for repos. |
| o1 | OpenAI | 200K tokens | 2024 | ~150K words | 100,000 out | Reasoning model, extended chain-of-thought. |
| o3-mini | OpenAI | 200K tokens | 2025 | ~150K words | 100,000 out | Cost-efficient reasoning model. |
| Claude Opus 4.7 | Anthropic | 200K tokens 1M tokens beta | 2026 | ~150K words | 32,000 out | Frontier model. 1M-token beta context for select customers. |
| Claude Sonnet 4.6 | Anthropic | 200K tokens 1M tokens beta | 2025 | ~150K words | 32,000 out | Balanced performance and price. 1M beta context available. |
| Claude Haiku 4.5 | Anthropic | 200K tokens | 2025 | ~150K words | 8,192 out | Fast, low-cost tier with full 200K context. |
| Grok 2 | xAI | 131K tokens | 2024 | ~98K words | Not published | Available via X Premium and the xAI API. |
| GPT-4o | OpenAI | 128K tokens | 2024 | ~96K words | 16,384 out | Multimodal flagship: text, image, audio input. |
| GPT-4 Turbo | OpenAI | 128K tokens | 2023 | ~96K words | 4,096 out | Predecessor to GPT-4o, still widely available. |
| Llama 3.1 405B | Meta | 128K tokens | 2024 | ~96K words | Not published | Open-weights flagship. Self-hosting possible. |
| Llama 3.3 70B | Meta | 128K tokens | 2024 | ~96K words | Not published | Smaller, faster open-weights model. Cheaper to host. |
| Mistral Large 2 | Mistral | 128K tokens | 2024 | ~96K words | Not published | European frontier model with strong multilingual support. |
| Command R+ | Cohere | 128K tokens | 2024 | ~96K words | Not published | RAG-optimized with native citations and tool use. |
| DeepSeek V3 | DeepSeek | 128K tokens | 2024 | ~96K words | Not published | Open-weights MoE model, strong reasoning at low cost. |