Long Context Availability
Sources: 1 • Confidence: Medium • Updated: 2026-04-13 03:49
Key takeaways
- A 1M-context window is generally available for the Opus 4.6 and Sonnet 4.6 models.
- Standard pricing applies across the full 1M-context window for Opus 4.6 and Sonnet 4.6, with no long-context premium.
- OpenAI and Google Gemini charge higher prompt prices once token counts exceed thresholds, including 200,000 tokens for Gemini 3.1 Pro and 272,000 tokens for GPT-5.4.
Sections
Long Context Availability
- A 1M-context window is generally available for the Opus 4.6 and Sonnet 4.6 models.
Long Context Pricing Mechanism
- Standard pricing applies across the full 1M-context window for Opus 4.6 and Sonnet 4.6, with no long-context premium.
Competitor Threshold Based Premiums
- OpenAI and Google Gemini charge higher prompt prices once token counts exceed thresholds, including 200,000 tokens for Gemini 3.1 Pro and 272,000 tokens for GPT-5.4.
Unknowns
- What official documentation (API limits, product pages, or release notes) confirms 1M context general availability for the two named models, and are there any eligibility constraints?
- What are the operational constraints at 1M context (rate limits, maximum request size in bytes, timeout limits, streaming behavior, and latency expectations)?
- How billing behaves in practice for near-1M prompts (invoice line items, rounding, any minimums), and whether any hidden multipliers apply at high token counts despite “standard pricing” wording?
- Are the competitor long-context pricing thresholds and step-up pricing mechanisms accurately represented, and what are the exact multipliers or tier prices?
- What decision-readthrough (operator/product/investor) is warranted from this corpus beyond “verify availability and pricing in official docs”?