Back to reviews
Gemini 3.1 Flash TTS

Gemini 3.1 Flash TTS

Google's new TTS API: 70 languages, 200+ audio tags, native multi-speaker

Gemini 3.1 Flash TTS is Google's new text-to-speech model, launched today on Google AI Studio and Vertex AI. It supports 70+ languages and introduces a natural-language audio tag system with 200+ expressivity controls — developers can describe delivery in plain English ("whisper conspiratorially", "warm and unhurried") and the model interprets those instructions at inference time. The model also supports native multi-speaker dialogue generation from a single prompt, outputting a conversation with distinct, consistent voices without requiring separate passes. All audio output is watermarked via Google's SynthID technology for provenance tracking. For developers building voice agents, podcasting tools, or multilingual apps, this is a meaningful upgrade over existing options. The audio tags approach in particular is a genuinely novel paradigm compared to prosody markup languages like SSML, and developer reception on X and HN has been strong — Simon Willison called out the expressivity controls as the standout feature.

Panel Reviews

The Builder

The Builder

Developer Perspective

Ship

This replaces ElevenLabs for a lot of use cases — and at Google's pricing it's hard to argue against. The natural-language audio tags are the real unlock: instead of wrestling with SSML prosody markup, you just describe what you want. The multi-speaker output from a single prompt is going to save a ton of orchestration code in voice agent pipelines.

The Skeptic

The Skeptic

Reality Check

Skip

It's Google — which means it could be deprecated in 18 months and replaced with Gemini 4 Flash TTS Pro Ultra. The audio tags sound creative but until there's a published spec for all 200+ of them, you're guessing at prompt-engineering your voice model. And SynthID watermarking is only as useful as the detection ecosystem, which is still nascent.

The Futurist

The Futurist

Big Picture

Ship

Natural-language expressivity control for TTS is a paradigm shift. When the model can interpret 'sound like you're delivering devastating news gently' without explicit prosody markup, we're entering an era where voice synthesis becomes genuinely directorial. The 70-language coverage plus SynthID watermarking points toward a future where synthesized voice is both globally expressive and auditably provenance-tracked.

The Creator

The Creator

Content & Design

Ship

I've been paying for ElevenLabs and manually tweaking prosody to get the right delivery. The audio tag system here could cut that iteration time dramatically — describing the scene and letting the model interpret is so much more intuitive than sliders and SSML. Multi-speaker from a single prompt is going to be huge for podcast generators and explainer video tools.

Community Sentiment

Overall790 mentions
72% positive20% neutral8% negative
Hacker News180 mentions
72%20%8%

Natural-language audio tags vs. SSML

Reddit210 mentions
68%22%10%

Comparison to ElevenLabs pricing

Twitter/X400 mentions
75%18%7%

Multi-speaker single-prompt generation