Gemini 3 Preview (OpenAI-compatible) rejects reasoning_effort: "medium" — works with low and high

Starting today, requests to Gemini 3 Preview using the OpenAI compatibility interface fail when reasoning_effort (thinking level) is set to "medium". The same request succeeds with "low" and "high".

This appears to be a model capability/config regression or a validation bug specific to the "medium" level.

Product / API: Gemini 3 Preview via OpenAI compatibility layer (Chat Completions API)

Issue type: Regression / breaking change

Note: There is no issue for gemini-3-flash-preview

Date first observed: 2025-12-19 (worked until 2025-12-16)


Steps to reproduce

  1. Send a request to Gemini 3 Preview (OpenAI compatibility endpoint) with:

    • model: "gemini-3-pro-preview" (or your exact Gemini 3 preview model id)

    • reasoning_effort: "medium"

    • Any basic prompt/messages

  2. Observe the error response.

  3. Repeat the same request with reasoning_effort: "low" → succeeds.

  4. Repeat with reasoning_effort: "high" → succeeds.


Minimal repro (example payload)

POST https://generativelanguage.googleapis.com/v1beta/openai/v1/chat/completions

{
  "model": "gemini-3-pro-preview",
  "reasoning_effort": "medium",
  "messages": [
    { "role": "user", "content": "Say hello." }
  ]
}


Actual result

Request fails with:

Thinking level MEDIUM is not supported for this model. Please retry with other thinking level

Hi @csekas
Thank you for reaching out!

Yes you are correct. I have tasted on my end using “reasoning_effort”: “medium” and encountering the same error. According to the Gemini documentation, the Pro model supports only low and high reasoning levels, whereas the Flash model supports low, medium, and high.
I recommend please go through this document, as it provides clear and more detailed information.

This looks like a bug presented as a feature. I mean, the Gemini API has a thinkingBudget field that accepts a numeric value. So how is it possible not to support the medium thinking level with an 8,192-token thinking budget?

In any case, thank you for taking the time, @Shivam_Singh2.

The problem is that each models has a different level of adherence to these kind of parameters. Initially, we wanted to provide full flexibility via the thinking budget, but it didn’t work super consistently before. So we opted to go with thinking levels. In the case of Pro, we did a bunch of eval work and medium didn’t produce good results nor was it consistent. I agree in general that this is basically a model bug at the moment. Hoping it gets fixed in the next rev of the model.

1 Like

So this is a model constraint, not of the API.
Make sence! Thank you Logan.