400, 500 and 503 since morning

Hi bro, unfortunately the error occurred again in the last 24 hours.

This time it’s not because the model I chose has reached the rate limit according to the plan.

  1. Model: gemini-2.0-flash, gemini-2.0-flash-lite
  2. Tier 1. And I have just upgraded to tier 2
  3. Error: [429 Too Many Requests] You exceeded your current quota, please check your plan and billing details

When I looked at the quota table, I found that there seemed to be an error in Google’s calculation. Perhaps some requests were counted in the free tier and the 429 errors belonged to this group. As shown in the table, I am in tier 1, but there is an error reaching the free tier’s rate limit.

It looks like you’re experiencing issues with the OpenAI API, specifically with token limits and server errors. Let’s break it down:

*Token Limit Issue:*

- The 400 error indicates that your input token count exceeds the maximum allowed limit of 131k tokens. Even as a paying customer, each model has a specific token limit.

- To resolve this, consider ¹ ²:

- \*Truncating or splitting input\*: Ensure your prompt and completion tokens don't exceed the model's context length. You can use utility libraries like \`tiktoken\` to count tokens.

- \*Optimizing prompt length\*: Review your prompts to minimize unnecessary tokens and reduce the input size.

*Server Errors (500/503):*

- These errors usually indicate issues on OpenAI’s end, such as server overload, maintenance, or internal timeouts.

- Possible causes and solutions ¹ ³ ⁴:

- \*Temporary server issues\*: Retry your request after a brief wait. If the issue persists, contact OpenAI support.

- \*Rate limiting or quota issues\*: Check your API usage dashboard for rate limits and quota exceeded errors. Consider implementing exponential backoff for retries.

- \*Model or endpoint issues\*: Verify the model name and endpoint are correct. Check the OpenAI status page for reported outages.

To further troubleshoot, consider:

- *Checking OpenAI’s status page*: Ensure there are no reported outages or maintenance scheduled.

- *Implementing error handling*: Use retry mechanisms with exponential backoff and circuit breakers to handle transient errors.

- *Contacting OpenAI support*: If issues persist, reach out to OpenAI support with your request ID and error details.The issue you’re experiencing with Gemini 2.5 Pro seems to be related to token limits. When you make requests to the Gemini API, it counts the total number of tokens in your input, including the prompt, context, and any files or data being processed. If this total exceeds the allowed limit of 1,048,576 tokens, you’ll get a 400 error.

Here are some potential solutions to help you resolve this issue:

- *Optimize Your Prompts*: Try to minimize unnecessary tokens in your prompts and reduce the input size. You can use utility libraries like `tiktoken` to count tokens.

- *Delete Unnecessary Files*: If your project includes many files or embedded data, consider deleting unnecessary ones or creating a `.aiexclude` file to prevent Gemini from loading heavy folders like `node_modules/`, `.next/`, or `dist/`.

- *Check Your API Calls*: Verify that your API calls are correctly formatted and that you’re not exceeding the model’s rate limits. You can check the API reference for request format, examples, and supported versions.

- *Upgrade Your Subscription Tier*: If your application needs exceed the current rate limits, consider upgrading to a higher subscription tier. Higher tiers offer increased limits, accommodating more extensive usage.

- *Request a Quota Increase*: If upgrading isn’t feasible, you can request a quota increase through the Google Cloud Console ¹ ² ³.

It’s also worth noting that Gemini 2.5 Pro and Gemini 2.5 Flash might have different token limits and usage policies. If you’re experiencing issues with Gemini 2.5 Pro but not with Gemini 2.5 Flash, it could be related to the specific limits or settings for eacIt seems like you’re experiencing issues with the Gemini API and AI Studio, but good news is that Google has restored access to the Gemini 2.5 Pro API through Google AI Studio, offering generous rate limits. Here’s what you need to know ¹ ²:

- *Gemini 2.5 Pro API Access*: You can access the API for free with limits of 60 requests per minute and 300,000 tokens per day. To get started, visit Google AI Studio, create an API key, and store it securely as an environment variable.

- *Optimizing Token Usage*: To avoid hitting token limits, consider optimizing your prompts, using structured system messages, and JSON schema hints to reduce output length. You can also leverage Gemini 2.5 Flash as a cost-saving alternative for certain tasks.

- *Troubleshooting*: If issues persist, ensure you’ve checked the API reference for request formats, and consider implementing error handling with retry mechanisms.

- *Additional Support*: For more complex projects or higher usage needs, explore upgrading to Vertex AI or contacting Google support for assistance.

Some key features of Gemini 2.5 Pro include ¹ ³:

- *Massive Context Window*: Supports up to 1 million tokens, enabling extensive document processing.

- *Enhanced Reasoning*: Incorporates simulated reasoning to validate outputs and reduce errors.

- *Multimodal Capabilities*: Supports text, images, audio, video, and code.

If you’re still facing issues or have questions, feel fIt looks like you’re getting a 400 error with Gemini 2.5 Pro due to exceeding the maximum token limit of 131,072, while Gemini 2.5 Flash works fine. Here are some potential reasons and solutions:

- *Token Limit*: Gemini 2.5 Pro has a specific token limit, and it seems like your input is exceeding that. You might need to optimize your prompts or reduce the input size to fit within the limit.

- *API Key or Project Issues*: Since creating a different API key or project didn’t work, it’s unlikely to be an issue with the key or project itself. However, you can try checking the API dashboard for any specific settings or restrictions.

- *Paid Account*: Having a paid account is great, but it’s possible that there are specific limits or settings applied to your account that are causing the issue. You might want to reach out to Google support to see if there’s something specific on their end.

It’s interesting that Gemini 2.5 Flash works fine, which suggests that the issue might be specific to the Pro version. If you’re still having trouble, consider asking on the Google AI Developers Forum for more specific guiIt sounds like you’re experiencing a frustrating issue with the Gemini API, specifically with a 400 error due to token limits on a paid plan. Given the recent changes and issues reported on the Google AI Developers Forum, it’s possible that the problem is related to the API itself.

Some users have reported similar issues with token limits and errors, even on paid plans. You might want to try optimizing your prompts or reducing the input size to see if that resolves the issue.

Additionally, since you’ve had issues with Google Cloud projects disappearing and needing to recreate them, it might be worth reaching out to Google support to see if there’s an underlying issue affecting your account or projects. Have you tried contacting support or checking the API status page for any known issues?dance.ree to ask!h model ¹.

Maaf saya saya tidak begitu handal dalam menyusun tapi jawaban Anda ada . Maaf bahasa ini bahasa saya dan saya mencintai bahasa saya

To give more context, my specific issue is that I’m on the free tier. My API key shows as ‘unavailable’ and I immediately get hit with a 429 error right from the start.

This makes me wonder: Is there no more free tier for the Gemini API? Has there been a recent policy change that removed it? It seems like new keys are being blocked or rate-limited instantly, even with no prior usage.

Has anyone heard anything about this?

Running into the same issue with gemini-2.5-flash
google.genai.errors.ClientError: 400 INVALID_ARGUMENT. {‘error’: {‘code’: 400, ‘message’: ‘Unable to submit request because the input token count is 131335 but model only supports up to 131072. Reduce the input token count and try again. You can also use the CountTokens API to calculate prompt token count and billable characters. Learn more: ``https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models’``, ‘status’: ‘INVALID_ARGUMENT’}}

I’m using a paid account via vertexai

The weird thing is that on some prompts that are 131335 I don’t get this issue. While on other prompts that are smaller, I get this issue consistently.

Gemini Code Assist 2.5 Pro facing this error for the past two days.

Agent Error, unknown agent message: [{ “error”: { “code”: 400, “message”: “The input token count (151486) exceeds the maximum number of tokens allowed (131072).”, “errors”: [ { “message”: “The input token count (151486) exceeds the maximum number of tokens allowed (131072).”, “domain”: “global”, “reason”: “badRequest” } ], “status”: “INVALID_ARGUMENT” } } ]