ComfyUI-OmniVoice-TTS
A new diffusion-based TTS model generates high-quality speech in over 600 languages with superior inference speed.
K2-FSA has launched OmniVoice, a groundbreaking text-to-speech (TTS) model that sets a new standard for multilingual speech synthesis. Built on a novel diffusion language model architecture, it supports an unprecedented 600+ languages in a zero-shot manner, meaning it can generate speech in languages it wasn't explicitly trained on for each voice. The model is designed for high-quality output with superior inference speed, and it comes with powerful features like voice cloning and voice design, allowing for precise control over vocal characteristics.
OmniVoice is now accessible through multiple platforms, including its official GitHub repository and a dedicated HuggingFace space for easy experimentation. Crucially, the model has been integrated into ComfyUI, a popular node-based interface for AI workflows, via the 'ComfyUI-OmniVoice-TTS' custom node. This integration allows users to seamlessly incorporate state-of-the-art, multilingual TTS into their visual AI pipelines for video generation, audiobook creation, or interactive applications, significantly lowering the barrier to entry for professional-grade speech synthesis.
- Supports over 600 languages with zero-shot capability, eliminating the need for per-language training data.
- Built on a novel diffusion model architecture for high-quality speech and superior inference speed compared to prior models.
- Enables voice cloning and design, and is integrated into ComfyUI for use in visual AI workflows.
Why It Matters
Dramatically lowers the cost and complexity of creating professional, multilingual audio content for global media, education, and accessibility tools.