Has anyone have been having issues with the Gemini live api multimodal input when it comes to streaming video live? The model seems to not know what’s in the image anymore for example I ask what color shirt I am wearing it says blue instead of grey. I also tested this with the boiler plate code google provides and same-thing.
Today video mostly doesn’t work for me. The model will hallucinate if pressed but can’t see anything.
Yeaa it started late last night for me, I’m wondering if it’s because they’re updating something.
We are looking for facts strong text
Continued to mess around with this and for some reason it detects images when using audio input and not text input…weird
Facing the same issue
Yep, big problem. Not sure what has changed
Hey folks, Is it hallucination or API not woking at all?? Are you still facing the same issue?? I just tested it’s working fine for me. This might be an intermittent issue due downtime or migration work. Let me if the issue persists.
Thanks
It is working now this problem was happening right before they changed from the experimental live api version to the current one. So I am assuming that was the reason it was acting up before