Gemini 2.5 Flash Thinking Tokens using OpenAI API

Hi,

Is it possible to enable/disable thinking or set thinking_tokens for Gemini 2.5 Flash using the OpenAI compatible API?

Thanks

8 Likes

I’m also interested in this question regarding controlling thinking with the OpenAI compatible API.

1 Like

waitting for updates

1 Like

It would be preferable for the model name to have two endpoints: one with Thinking and one without Thinking, which would make integration easier. A bit like what OpenRouter currently offers.

or like claude, add extra info, more control

1 Like

Maybe even like the ones in Requesty? Where you can set one of four thinking efforts:

Same here. I would like to see OpenAI compatibility with Thinking Tokens and Reasoning Trace.

Same as well, I want to use and disable thinking mode.

1 Like