Context Window #186494
Replies: 12 comments 2 replies
-
|
Beta Was this translation helpful? Give feedback.
-
|
Claude Opus 4.6 “supporting 1M context” is basically the model’s maximum architectural capability, but in the actual product Anthropic sets practical limits for speed, cost, and reliability. At 1M tokens the compute and latency blow up massively, and even if the model can technically read that much, its effective recall and consistency can become less stable. That’s why they cap it at 128K input and 64K output to keep performance predictable, inference fast enough, and pricing manageable. |
Beta Was this translation helpful? Give feedback.
-
|
Claude Opus 4.6 can support up to ~1M tokens at the model level. Copilot Chat does not expose the model’s maximum context. Copilot applies its own caps for:
So you see: 128K input / 64K output Important distinction Model capability ≠ Product limit. Vendors frequently gate large contexts behind:
Copilot currently uses a constrained configuration of Claude models. Why Copilot does this
GitHub optimizes for interactive developer workflows, not massive document ingestion. If you truly need >128K Use Claude directly through Anthropic’s API or a provider that exposes 1M context. Copilot is not the right surface today. |
Beta Was this translation helpful? Give feedback.
-
|
Claude Opus and Sonnet 4.6 both just changed 1M context window from beta to GA. Copilot should make use of it as well |
Beta Was this translation helpful? Give feedback.
-
|
When this will be available in GitHub Copilot? Or we need to migrate to Claude Code? |
Beta Was this translation helpful? Give feedback.
-
|
pretty disappointed with context window for Claude opus 4.6 in copilot.... its not even usable as it easily reaches limit .... compared to open and gemini modals have huge context windows |
Beta Was this translation helpful? Give feedback.
-
|
Over the weekend got to play with Claude Code and the 1M token window. The difference is night and day. Lets you do some really complex work without having to worry about the main agent getting anywhere near the context window. Keeps long running agentic work on track without the risk of compaction causing it to go off track. We really need this in copilot asap. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
|
How about upgrading to 400k?I tested GPT5.4-400k and found that context compression reduces some of it in my workspace. Of course, it may not be possible to make such a simple comparison. |
Beta Was this translation helpful? Give feedback.
-
|
Maybe 1M is too big, but 200k is too small. Middle ground would be 400k for me. Letting it work on a task without "compacting conversation and ai model getting confused right after and messing the whole progress up..".. well that would be great. |
Beta Was this translation helpful? Give feedback.
-
|
With such a small context window, Opus looks like a complete idiot when handling complex projects. |
Beta Was this translation helpful? Give feedback.
-
|
GitHub Copilot sets its own context limits independent of what the underlying model supports. The 128K input / 64K output cap on Opus 4.6 is a GitHub product decision, not something Anthropic controls. Why GitHub does this: running 1M token inference at scale is expensive and significantly slower. Interactive coding workflows are latency-sensitive, and serving 1M contexts across a large user base with sub-second response expectations doesn't work economically right now. So GitHub capped it at something usable for the product. The 1M context window did go GA on the Anthropic side in early 2026, but GitHub hasn't updated Copilot's limits to reflect that yet. There's no public timeline for when (or if) they'll increase it. If you need the full 1M window, your options are Claude Code or the Anthropic API directly. Copilot is not the right surface for that use case at the moment. |
Beta Was this translation helpful? Give feedback.


Uh oh!
There was an error while loading. Please reload this page.
-
Select Topic Area
Question
Copilot Feature Area
VS Code
Body
Why does the Claude Opus4.6 token context window only have 128K input and 64K output, when the model can handle up to 1M?

Beta Was this translation helpful? Give feedback.
All reactions