I am consistently encountering fatal agent terminations when using Claude Opus 4.6 (Thinking) on Antigravity IDE. Instead of managing context gracefully, the agent simply terminates midway through any moderately complex session.
Here is the exact Debug Info I keep seeing:
-
Error:
HTTP 400 Bad Request -
Detail:
prompt is too long: 218849 tokens > 200000 maximum
The Core Architectural Defect (NOT a User Error): Antigravity’s current “smarter context packaging” is fundamentally broken when handling long-context reasoning models. The IDE completely lacks a dynamic auto-compaction or context pruning mechanism.
Instead of intelligently summarizing the trajectory when approaching the 200K threshold, it recklessly dumps the entire accumulated history alongside the massive, unpredictable “Thinking” token overhead generated by Opus 4.6 straight into the API payload. This guarantees an instant token overflow and a fatal crash.
The Industry Standard (Why Antigravity is Falling Behind): This is already a solved problem in the industry.
-
Cursor IDE handles this flawlessly with background context summarization.
-
Anthropic’s own Claude Code utilizes robust native compaction logic (e.g., the
/compactfunction).
Why is a flagship, “agent-first” IDE from Google so far behind, lacking even the most basic auto-compaction feature?
As long-context reasoning models like Opus 4.6 become the new standard, relying on hard-coded token limits without intelligent history summarization is fundamentally unsustainable. Please prioritize implementing a native Server-side Compaction mechanism immediately. We are tired of losing our entire task trajectories to a preventable overflow.
