VoxCPM2
Describe a voice in text, get studio-quality speech — no reference audio needed
VoxCPM2 is a 2B-parameter text-to-speech system from OpenBMB — the team behind MiniCPM — built around a tokenizer-free, diffusion-autoregressive architecture. Most TTS systems convert text to discrete audio tokens first, then decode those tokens to waveform. VoxCPM2 skips the tokenization step entirely, operating in continuous latent space. The result is 48kHz output with smoother prosody and finer pitch control than token-based systems. The headline feature is "Voice Design": you describe a voice in natural language — "a confident male voice, mid-Atlantic accent, slightly gravelly, deliberate pacing" — and VoxCPM2 synthesizes a brand-new voice from that description without any reference audio sample. This is architecturally different from voice cloning (which requires samples) and voice selection (which picks from a catalog). It supports 30 languages with automatic detection, no language tags required. The model runs on consumer hardware (~8GB VRAM), integrates with the MiniCPM-4 language model backbone, and is released under Apache 2.0. For developers building multilingual voice products or researchers exploring generative voice control, VoxCPM2 represents a meaningful step beyond current open TTS leaders like F5-TTS and CosyVoice.
Panel Reviews
The Builder
Developer Perspective
“The tokenizer-free architecture is the right technical move — eliminating the quantization artifacts from discrete audio tokens is the main reason commercial TTS still sounds better than open source. The Voice Design feature alone is worth experimenting with for anyone building voice products. 8GB VRAM requirement is very reasonable.”
The Skeptic
Reality Check
“48kHz is great on paper, but the diffusion-based approach likely trades inference speed for quality. No benchmarks are published against F5-TTS or Kokoro in the README, which is a red flag. Voice Design sounds novel but natural-language voice descriptions are inherently ambiguous — you'll get inconsistent results across generations.”
The Futurist
Big Picture
“Voice Design as a primitive changes how voice AI gets built. Instead of recording actors, teams can describe and iterate on synthetic voices the way designers iterate on color palettes. When this technology matures, every product that uses voice will have a unique, consistent, describable brand voice — not a voice cloned from someone else.”
The Creator
Content & Design
“Finally a TTS tool where I can describe what I want instead of auditioning samples. For narration, podcasts, and video, being able to say 'warm, unhurried, slightly husky' and get a consistent voice is a workflow unlock. The 30-language automatic detection is huge for multilingual content creators — no more manually tagging each segment.”
Community Sentiment
“Voice Design feature and tokenizer-free approach”
“Comparison requests against F5-TTS and Kokoro”
“Voice Design demos and 30-language support”