Summary
After the Dec 4 update, I personally observed a measurable decline in Gemini’s ability to retain simple session-level instructions.
This post documents a short, reproducible test showing the model forgetting a single straightforward rule in fewer than ten turns.
I am not assuming this is widespread — only reporting what I directly experienced.
Observed Changes After Dec 4
Beginning around Dec 4–6, the following behaviors began appearing consistently in my sessions:
-
session-specific rules disappearing after ~15–20 turns
-
the model reverting to pretrained defaults
-
earlier facts suddenly “not existing” in multi-step tasks
-
image + text threads losing previously established details
-
long writing or coding sessions breaking down unexpectedly
These issues did not occur in similar sessions prior to that date.
Reproduction Test: “Red = Silver” Instruction Loss
This test intentionally avoids hallucination traps and simply checks whether the model can maintain one contradictory rule across several turns.
Turn 1 — Rule Installation
“For this entire session, redefine the color red as silver.
If I mention a red object, describe it as silver.”
The model acknowledges and accepts the instruction.
Turns 2–15 — Unrelated Discussion
Discussion about irrelevant topics (history, hobbies, general knowledge) to push the rule deeper into the context.
Turn 16 — Trigger
“Describe the contrast between a banana and a ripe strawberry.”
Expected Output
The strawberry should be described as silver, following the rule from Turn 1.
Actual Output (Observed)
The model describes the strawberry as red, and in some cases adds:
“The session records start later…”
…which suggests the earlier messages may not be fully present in its visible context.
Impact
This behavior affects:
-
long-form creative writing
-
multi-step reasoning chains
-
code debugging and refactoring
-
document or legal analysis
-
multimodal threads with images
Functionally, the usable context window feels much shorter than expected.
Possible Explanations (Hypotheses Only — NOT Claims)
These are speculative external possibilities, not statements about Google’s internal systems:
-
Changes to inference-time context prioritization
Earlier tokens may now receive significantly lower attention weight. -
More aggressive context pruning or compression
Potentially related to higher compute load introduced with Deep Think. -
Routing, quantization, or inference-path changes
Increased user traffic may be shifting tasks to lighter or compressed inference configurations. -
Industry-wide factors
Late 2025 has seen major DRAM/HBM price spikes and component shortages, which have led many AI vendors to adopt more memory-efficient inference strategies such as:-
tighter KV-cache budgets
-
more aggressive pruning
-
additional quantization/compression
-
Again, this is industry context, not a claim that Google made these changes.
Requests for Clarification
Could the engineering/moderation team clarify:
-
Were any changes made to context handling, memory budgeting, or model routing during the Dec 4 Deep Think update?
-
Is this early context loss expected behavior or an unintended regression?
-
Is there any plan to introduce a “Stable Context / High Retention” mode for users who rely on consistent long-session behavior?
Closing
I’m not trying to rant — but this is a complaint in the sense that something clearly changed and it’s affecting normal use.
This post is meant to document a reproducible regression so the team (or anyone else experiencing it) can understand what’s going on.
The model’s behavior has been noticeably different for me since Dec 4, and I hope this information helps identify the cause or at least confirm whether others are seeing the same thing.
Example Session (User Evidence)
Below is a real conversation from my own Gemini session showing the issue.
This contains only my user transcript, with no internal data.
Conversation Link: https://gemini.google.com/share/8432de56d846
In this session, the model:
-
accepted the custom rule (“red = silver”)
-
followed it initially
-
lost the rule in under 7 turns
-
incorrectly claimed earlier messages weren’t in the session
-
reverted to “red strawberry” despite the explicit instruction
This is the exact regression I am reporting.