Anthropic made a significant change to Claude's 1 million token context window this week — and if you use Claude at any level, it's worth understanding what happened. The short version: the pricing penalty for sending Claude a really long prompt is gone. Here's what that means, who it affects, and whether you need to do anything about it.

What Is a Token? (And Why Claude's Context Window Matters)

Before anything else, let's demystify the jargon. A token is roughly three-quarters of a word. So when you hear "1 million tokens," think approximately 750,000 words — that's the entire Lord of the Rings trilogy, or a medium-sized codebase. A context window is basically how much information Claude can hold in its head at once during a single conversation.

The bigger the context window, the more Claude can see, reference, and reason across — without you having to repeat yourself or break things up into chunks.

Claude's 1 Million Token Context Window: What Anthropic Changed

Anthropic announced that its two flagship models — Claude Opus 4.6 and Claude Sonnet 4.6 — now support a 1 million token context window at standard pricing. No surcharge. No penalty for going big.

Here's the thing that matters: the models technically supported large contexts before. But once your prompt crossed roughly 200,000 tokens, you got bumped into a premium pricing tier. The whole request became more expensive — sometimes double the rate.

That pricing distinction is now gone. A 900,000-token request is billed at the same per-token rate as a 9,000-token one. claude

Claude Opus 4.6 and Sonnet 4.6 Pricing: What You'll Pay Now

Standard pricing is $5 per million input tokens and $25 per million output tokens for Opus 4.6, and $3 per million input tokens and $15 per million output tokens for Sonnet 4.6. claude Previously, once you crossed that long-context threshold, input pricing could double. That penalty is now gone entirely.

The Other Change Nobody's Talking About: 6x More Media Per Request

The pricing change got most of the headlines, but there's another upgrade baked into this announcement: media limits have expanded to 600 images or PDF pages per request, up from 100. claude If you regularly feed Claude large documents with embedded images — think contracts, research papers, or pitch decks — that's a meaningful improvement.

Why Developers Built Workarounds — And Why They May Not Need To Anymore

Because cost used to make large prompts risky. Developers built entire systems specifically to avoid sending too much information to Claude at once. Instead, they'd use "retrieval" systems that searched for and pulled only the most relevant snippets before sending anything to the model.

This worked, but added complexity, extra engineering time, and still sometimes caused Claude to miss important context that lived outside those snippets.

With the premium tier gone, developers now have a genuine choice between those two approaches — based on performance and preference, not just cost.

Does Claude Perform Well at 1 Million Tokens?

A bigger context window only matters if the model can actually use it well. Anthropic addressed this directly: Opus 4.6 scores 78.3% on MRCR v2, which they describe as the highest among frontier models at that context length. claude MRCR is a benchmark that tests how accurately a model can retrieve and reason across information buried deep in a long context — it's the difference between a model that technically "accepts" a million tokens and one that actually uses them well.

Real companies are already seeing the difference. One engineering team reported a 15% decrease in compaction events claude — meaning Claude needed to forget or summarize earlier parts of the conversation far less often. Another developer noted that previously, large code reviews had to be broken into chunks, causing Claude to lose track of cross-file dependencies. With the larger window, the full review goes in at once and quality improves.

Who Can Access Claude's 1 Million Token Context Window?

1M context is available on the Claude Platform natively and through Amazon Bedrock, Google Cloud's Vertex AI, and Microsoft Foundry. Claude Code Max, Team, and Enterprise users on Opus 4.6 get the full 1M context window by default. claude

The Catch: Bigger Context Still Costs More Tokens

"No surcharge" doesn't mean "free." Token usage still adds up with input size, and a fully maxed-out Sonnet prompt still costs real money. The real shift is in the experimentation barrier — you can now test long-context workflows without fear of accidentally triggering a premium rate. Whether it makes sense for your production setup is a separate question.

Frequently asked questions

What is a context window in AI?

A context window is essentially how much information an AI model can hold in its memory at one time. Everything you type, every file you attach, and every response the AI gives all share that same pool of space. Once you hit the limit, the model either starts forgetting earlier parts of the conversation or can't continue cleanly. A bigger context window means Claude can handle longer documents, larger codebases, and more complex conversations without losing track.

Does Anthropic's 1 million token context window affect Claude Pro users?

Mostly not directly — at least not in terms of billing. The pricing change is aimed at developers and businesses using Claude through the API. If you use Claude on a Pro, Team, or Enterprise plan, you pay a flat monthly rate rather than per token. The context window improvements do apply to you though, meaning Claude can hold more of your conversation in memory without "forgetting" earlier parts of long sessions.

What about Claude Code users?

This one's a direct win. Claude Code Max, Team, and Enterprise users running Opus 4.6 now get the full 1M context window automatically, with no extra usage required. That means longer coding sessions, fewer interruptions, and Claude staying aware of more of your codebase at once.

What's the practical difference between 200K and 1M tokens?

200,000 tokens is roughly 150,000 words — a long novel or a decent-sized codebase. 1,000,000 tokens is closer to 750,000 words — think multiple large codebases simultaneously, an entire legal case file, hundreds of research papers, or a full year of internal documents. The jump isn't just quantitative; it changes what kinds of tasks are even possible in a single session.

Do I have to do anything to enable this?

No. No beta header is required — requests over 200K tokens now work automatically. If you were already sending a beta header, it's ignored, so no code changes are required.

Can I send 600 images now?

Yes — that's new too. Media limits have expanded to 600 images or PDF pages per request, up from 100, available on Claude Platform natively, Microsoft Azure Foundry, and Google Cloud's Vertex AI.

Does Claude actually perform well at 1 million tokens, or does it get confused?

Legitimate question — bigger context doesn't automatically mean better performance. Anthropic points to Opus 4.6's 78.3% score on MRCR v2 as evidence that the model performs well even at the far end of its context window. That benchmark specifically tests whether a model can find and correctly use information buried deep in a long document — not just whether it technically accepted the input.

Is Claude getting cheaper in 2026?

In one meaningful way, yes. Anthropic eliminated the premium pricing tier that previously kicked in for large prompts. The base prices haven't dropped — Opus 4.6 is $5/$25 and Sonnet 4.6 is $3/$15 per million tokens — but removing that surcharge is a real cost reduction for anyone regularly working with large amounts of text or data.

Does this make RAG obsolete?

Not quite. Retrieval-augmented generation still makes sense for truly massive datasets that exceed even 1M tokens, or when speed and cost efficiency are the top priority. But for many teams, the complexity of building a retrieval system was partly a workaround for context limits and pricing penalties. With those reduced, some teams may choose to simplify their architecture.

Is Anthropic ahead of competitors on this?

Not exactly — this is more of a catch-up move. Google's Gemini and OpenAI already offered models with comparable context windows. What's notable is the removal of the pricing premium, which makes large-context usage more accessible for teams that were previously priced out of experimenting with it.

Tim Pettigrew

Tim Pettigrew

Pittsburgh Realtor · eXp Realty · Tim Sells Pittsburgh

I've been selling real estate in Pittsburgh since 2018 — mostly in the Alle-Kiski Valley, but these days you'll find me all over the city. When I'm not helping buyers and sellers navigate the market, I'm writing about AI tools, tech, and what it all means for regular people. No jargon, no hype — just straight talk. The Pittsburgh Pulse is my little corner of the internet for exactly that.

The Pittsburgh Pulse

Real estate + AI + Pittsburgh — delivered to your inbox.

Subscribe free

Keep Reading