This is a fairly common issue. Gemini 3.1 Pro, even with a small context (<32k), completely ignores prompts like “write an answer of X lines/X pages/X words/X characters/X tokens.” I don’t know if this is done to save money or not, but if I ask the model to give a long answer, I, as in the case of 3.0 pro and 2.5 pro, want to get a LONG and detailed answer.
I’m also extremely confused by the model’s inability to determine what I’m asking her to reflect on. I was recently conducting a small psychology study and asked a model to reflect on my past messages and provide a short answer about how Person A and Person B would react in the given situations. The request was for a straightforward answer, so I could copy and paste it as fact, not just a thought. The result: The model’s “Thought Process” only considered my request, not my past messages, which I asked for. This resulted in the model, instead of a short answer like “Person A will want to go to the movies, but Person B will refuse,” producing a huge, meaningless tranche of thoughts about why and how she arrived at this conclusion. In my opinion, this should have been in the “Thought Process,” but not in the answer itself, which was supposed to be brief.
I’d also like to point out that the model doesn’t respect the request for detailed reasoning. Despite the “High” mode enabled, it thinks very briefly, again dumping its thoughts into the main answer, which in some situations is unacceptable for me. 2.5 Pro was more flexible in this regard, as the number of tokens in its reasoning was quite configurable, which made it possible to force the model to truly consider its answer before giving it.
Therefore, I believe that the 3.1 Pro model is completely incapable of following instructions, and its “Thought Process” is also flawed.
Is this issue still relevant to anyone? If so, please write in the comments.
(I don’t speak English, but I hope I’ve made my thoughts clear.)