Gemini 3.1 Flash Live launches with 90.8% ComplexFuncBench audio score
Google says its new realtime voice model improves noisy-environment understanding, long conversations and function calling, and it's rolling into Gemini Live, Search Live and AI Studio. Voice creators can test it for lower-latency spoken interactions.

TL;DR
- Google's launch post introduces Gemini 3.1 Flash Live as a new realtime voice-and-vision model, with the company framing it as a step up in quality, reliability, and latency.
- In Google's DeepMind thread, the new model is described as better at handling noisy environments, longer conversations, and task completion through improved function calling.
- The rollout post says Gemini 3.1 Flash Live is going live in Gemini Live, Google Search Live, and Google AI Studio, so both end users and developers can access it now.
What changed for voice creators
Google is pitching this release as a practical upgrade for spoken interfaces, not just a benchmark refresh. The launch materials in the main announcement show Gemini 3.1 Flash Live hitting 90.8% on ComplexFuncBench audio and 95.9% on Big Bench Audio speech reasoning, with the included charts
and
positioning it ahead of earlier Gemini native-audio versions.
For creators building voice-led experiences, the more useful change is behavioral. Google's product thread says the model keeps track of longer conversations, understands details in noisy settings, and makes function calling more reliable, which maps directly to voice agents that need to listen, remember context, and trigger tools without awkward retries. The availability update says it is already rolling into Gemini Live and Search Live, while developers can start testing it in AI Studio and dig into the fuller product write-up via Google's overview.