I’m using Gemini-2.5-pro API but the response time is slow, it is taking around ~5mins to get a response mostly because of thinking + very long response. I’m using Vercel serverless functions for API but getting function timeout , how do i resolve this
I’ve never seen it take 5 minutes personally, but do remember 2.5 Pro is still pre-release and not for production use. Which means you have to contend with issues like this during this incubation period.
Going forward, if you need performance SLAs / reserved API capacity, then you can pay for that via Vertex AI.