Hi – the gemini-2.0-flash-live “family” of models has been great.
It would be super useful if, in addition to PCM16 i/o, these streaming models supported mulaw / g711_ulaw. Is that possible?
My application pipes data to/from a phone system and currently I have to re-encode to/from PCM16 on the fly. This works but it’s choppy. The OpenAI realtime system supports g711_ulaw i/o and it’s much smoother – but I’d rather stick with gemini