Which Open Source Text-to-Speech Model Should You Use?
A Practical Comparison of the Best Open-Source TTS Models in 2025

The demand for high-quality Text-to-Speech technology has grown dramatically in recent years, driven by applications ranging from accessibility tools and virtual assistants to content creation and multilingual communication. While proprietary solutions from major tech companies offer excellent quality, they come with significant limitations, including usage costs, vendor lock-in, data privacy concerns, and restricted customization capabilities.
Open-source TTS models have emerged as compelling alternatives, offering complete control over data and infrastructure, extensive customization possibilities, elimination of per-usage costs after deployment, and community-driven innovation. For developers, researchers, and organizations prioritizing data sovereignty or requiring specialized voice characteristics, open-source solutions represent the path forward for scalable, cost-effective speech synthesis.
This analysis evaluates five leading open-source TTS models through systematic benchmarking on the LibriTTS dataset, examining their performance across voice cloning accuracy, content fidelity, inference speed, and resource efficiency.
Model Overview
XTTS v2
XTTS v2, represents a significant advancement in multilingual zero-shot Text-to-Speech synthesis. Building upon the Tortoise model architecture, XTTS v2 employs a GPT2-based decoder that predicts audio tokens computed by a pre-trained Discrete Variational Autoencoder (VAE). The model's core innovation lies in its sophisticated speaker conditioning mechanism using a Perceiver architecture.
The Perceiver component processes mel-spectrograms from reference audio and produces 32 latent vectors that represent speaker-specific characteristics. These vectors are then prefixed to the GPT decoder, enabling the model to capture speaker characteristics more effectively than simpler encoder approaches used in models like Tortoise or speech prompting methods employed in VALL-E. This Perceiver-based conditioning provides consistent model outputs between different runs, alleviating the speaker shifting issues that plagued earlier implementations. The architecture supports multiple reference audio clips without length limitations, allowing for the capture of different aspects of target speakers or even the combination of characteristics from multiple speakers to create unique voices.
The training regimen for XTTS v2 involved approximately 2,354 hours of multilingual data, including 541.7 hours from LibriTTS-R and 1,812.7 hours from LibriLight for English alone. The model supports 16 languages, including English, Spanish, French, German, Italian, Portuguese, Polish, Turkish, Russian, Dutch, Czech, Arabic, Chinese, Hungarian, Korean, and Japanese.
IndexTTS
IndexTTS, builds upon the XTTS foundation but introduces several architectural improvements designed for industrial-level deployment. The model maintains the GPT-style transformer architecture but incorporates a conformer-based speech conditional encoder and replaces the original speech decoder with BigVGAN2 for enhanced audio quality and stability.
A key architectural advancement in IndexTTS is its exploration of Vector Quantization (VQ) versus Finite-Scalar Quantization (FSQ) for better codebook utilization in acoustic speech tokens. The FSQ approach demonstrates superior codebook utilization compared to traditional VQ methods, leading to more efficient representation of speech information. The model uses a simplified text processing pipeline, removing the front-end Grapheme-to-Phoneme (G2P) module and instead employing raw text input with a Byte-Pair Encoding (BPE) based tokenizer with a vocabulary size of 12,000 tokens.
The architecture follows a three-stage pipeline similar to XTTS: speech-to-codec VQVAE encoding, text-to-codec language modeling, and latent-to-audio decoding through BigVGAN2. The model was trained on tens of thousands of hours of data and demonstrates significant improvements in naturalness, content consistency, and zero-shot voice cloning compared to XTTS.
CosyVoice 2.0
CosyVoice 2.0, represents a fundamental shift in TTS architecture through its use of supervised semantic tokens. Unlike traditional approaches that rely on unsupervised speech tokenization, CosyVoice derives its speech representations from a multilingual automatic speech recognition model by inserting vector quantization layers into the encoder architecture.
The model employs a two-stage generation process: progressive semantic decoding using both Language Models (LMs) and Flow Matching. The supervised semantic tokens are extracted from a fine-tuned proprietary SenseVoice ASR model trained on multilingual audio data, with vector quantization inserted between encoder layers to obtain discrete representations. This approach provides explicit semantic information and better alignment between text and speech compared to unsupervised methods.
A significant architectural innovation in CosyVoice 2.0 is the introduction of Finite-Scalar Quantization (FSQ) to replace traditional Vector Quantization (VQ) in the speech tokenizer, leading to improved codebook utilization and better capture of speech information. The model architecture is streamlined by removing the text encoder and speaker embedding components, allowing pre-trained textual Large Language Models to serve as the backbone, enhancing context understanding capabilities.
The unified framework enables both streaming and non-streaming synthesis within a single model through a chunk-aware causal flow matching approach. This design achieves bidirectional streaming support with latency as low as 150ms, while maintaining high-quality audio output.
Fish Speech
Fish Speech, introduces a novel Dual Autoregressive (Dual-AR) architecture designed to address stability issues in Grouped Finite Scalar Vector Quantization (GFSQ) during sequence generation. The architecture consists of two sequential autoregressive transformer modules operating at different temporal scales.
The Slow Transformer processes input text embeddings to capture global linguistic structures and semantic content, operating at approximately 20 tokens per second. The Fast Transformer then processes the output from the Slow Transformer to generate detailed acoustic features. This serial fast-slow design improves the stability of GFSQ while maintaining high-quality output, making it particularly effective for AI interactions and voice cloning applications.
A crucial component of the Fish Speech system is the Firefly-GAN (FF-GAN) vocoder, which integrates multiple vector quantization techniques including Finite Scalar Quantization (FSQ) and Group Vector Quantization (GVQ). The FF-GAN architecture achieves 100% codebook utilization, representing state-of-the-art performance in vector quantization efficiency. The model eliminates traditional Grapheme-to-Phoneme conversion by leveraging Large Language Models for linguistic feature extraction, streamlining the synthesis pipeline and enhancing multilingual support.
The training corpus for Fish Speech is substantial, comprising over 720,000 hours of multilingual audio data, enabling the model to learn diverse linguistic patterns and pronunciation variations across eight supported languages.
F5-TTS
F5-TTS, implements a fully non-autoregressive approach based on Flow Matching with Diffusion Transformer (DiT) architecture. The model's design philosophy centers on simplicity, eliminating complex components such as duration models, text encoders, and phoneme alignment systems that characterize traditional TTS architectures.
The core architectural innovation involves padding text input with filler tokens to match the length of input speech, then performing denoising operations for speech generation. To address the slow convergence and low robustness issues observed in the original E2 TTS design, F5-TTS incorporates ConvNeXt blocks to refine text representation, facilitating better alignment with speech features.
A key contribution of F5-TTS is the Sway Sampling strategy applied during inference. This sampling strategy for flow steps significantly improves model performance and efficiency, and can be applied to existing flow-matching-based models without requiring retraining. The strategy optimizes the inference process, contributing to the model's impressive Real-Time Factor of 0.15.
The model architecture employs a Diffusion Transformer with U-Net style skip connections, trained on a text-guided speech-infilling task. The training involved approximately 100,000 hours of multilingual data, enabling the model to exhibit highly natural and expressive zero-shot capabilities with seamless code-switching functionality.
Model Specifications Summary
| Model | Architecture | Parameters | Key Innovation | Languages | Streaming | Training data |
XTTS v2 | GPT2 decoder with Perceiver conditioning | ~1.1B | Perceiver-based speaker conditioning | 16 languages | Yes (200ms) | 2,354 hours |
IndexTTS | GPT-style with conformer conditioning | ~50M (VAE)B | VQ vs FSQ analysis | Multilingual | No (adaptable) | 50,000+ hours |
CosyVoice 2.0 | Text-speech LM with Flow Matching | 0.5B | Supervised semantic tokens + FSQ | 5+ languages | Yes (150ms) | 10,000+ hours |
Fish Speech | Dual-AR (Slow/Fast Transformers) | 4B (S1), 0.5B (mini) | GFSQ 100% codebook | 8 languages | Limited | 720,000+ hours |
F5-TTS | Diffusion Transformer + ConvNeXt | ~200M | Sway Sampling strategy | Multilingual | No | 100,000 hours |
Evaluation Methodology
Dataset: LibriTTS
Our benchmark evaluation was conducted on the LibriTTS dataset, introduced by Zen et al. (2019) as a comprehensive corpus specifically designed for Text-to-Speech research. LibriTTS consists of approximately 585 hours of read English speech at a 24kHz sampling rate from 2,456 unique speakers, derived from LibriSpeech materials but optimized for TTS applications.
The dataset maintains gender balance and employs sentence-level segmentation rather than silence-based splitting, creating more natural linguistic units. It preserves both original and normalized text representations, enabling evaluation of models' ability to handle natural text formatting, including capitalization and punctuation. The corpus is organized into seven subsets following LibriSpeech's structure, with our evaluation utilizing the test-clean subset for high-quality, low-noise conditions.
We prepared 100 test cases from the LibriTTS test-clean subset, ensuring each selected speaker had at least two audio samples to enable proper reference-target pairing for voice cloning evaluation. This methodology provides sufficient data for statistical analysis while following established practices in zero-shot TTS evaluation.
Evaluation Protocol
All experiments were performed in batch mode on Google Colaboratory using their T4 GPU infrastructure to ensure consistent hardware conditions across all model evaluations. Each test case consists of a reference audio sample and corresponding text, paired with target text for synthesis. The reference audio serves as the voice template, while models generate new speech using the target text in the reference speaker's voice.
We assessed each model across multiple dimensions following established TTS evaluation practices. Speaker Similarity was measured using SpeakerRecognition from SpeechBrain with the ECAPA-VOXCELEB model, quantifying how well the synthesized speech matches the reference speaker's voice characteristics on a continuous scale where higher values indicate better voice preservation.
Word Error Rate measured content accuracy by comparing ASR transcriptions of generated audio against ground truth text using OpenAI's Whisper-base model. The WER calculation follows the standard formula: (Substitutions + Insertions + Deletions) divided by total words in the reference text. Lower WER values indicate better content fidelity and intelligibility.
Real-Time Factor calculated the ratio of synthesis time to audio duration, where values below 1.0 indicate faster-than-real-time generation. This metric is crucial for applications requiring interactive or streaming synthesis capabilities.
Benchmark Results
Overall Performance
| Model | Samples | Speaker Similarity | Word Error Rate | RTF | GPU Memory (MB) |
XTTS_v2 | 100 | 0.4973 | 0.2750 | 0.482 | 5,040 |
IndexTTS | 100 | 0.4978 | 0.2418 | 0.848 | 4,830 |
CosyVoice2-0.5B | 100 | 0.5881 | 0.3237 | 1.283 | 3,379 |
F5TTS | 100 | 0.4788 | 0.3204 | 1.283 | 2,994 |
FishSpeech-S1-mini | 100 | 0.5951 | 0.5448 | 31.467 | 4,029 |
Performance Rankings
| Rank | Speaker Similarity | Word Error Rate | Real-Time Factor | Latency | GPU Memory (MB) |
1st | CosyVoice (0.5881) | IndexTTS (0.2418) | XTTS_v2 (0.482) | XTTS_v2 (3.36s) | F5TTS (2,994 MB) |
2nd | FishSpeech (0.5951) | XTTS_v2 (0.2750) | IndexTTS (0.848) | F5TTS (3.44s) | CosyVoice (3,379 MB) |
3rd | XTTS_v2 (0.4973) | F5TTS (0.3204) | F5TTS (0.894) | IndexTTS (4.76s) | FishSpeech (4,029 MB) |
4th | IndexTTS (0.4978) | CosyVoice (0.3237) | CosyVoice (1.283) | CosyVoice (7.71s) | IndexTTS (4,830 MB) |
5th | F5TTS (0.4788) | FishSpeech (0.5448) | FishSpeech (31.467)) | FishSpeech (102.16s) | XTTS_v2 (5,040 MB) |
Streaming Support Analysis
The models vary significantly in their streaming capabilities. XTTS_v2 provides full streaming support with 200ms time to first chunk, making it excellent for real-time applications. CosyVoice 2.0 offers a unified streaming and non-streaming framework, providing flexibility for different use cases. Fish Speech has limited streaming capabilities in its current implementation. IndexTTS does not natively support streaming but can be adapted with additional engineering effort. F5TTS uses a non-autoregressive design that inherently limits streaming capabilities.
All models in our benchmark were evaluated in batch mode. While some models like IndexTTS don't natively support streaming, they can be modified to provide near-streaming capabilities through techniques like chunk-based processing and output buffering, though this requires additional development work and may impact performance.
Analysis and Discussion
The benchmark results reveal distinct strengths and weaknesses across the evaluated models. CosyVoice 2.0 and Fish Speech achieve the highest speaker similarity scores, demonstrating superior voice characteristic preservation. This makes them particularly suitable for applications requiring high-fidelity voice cloning. However, Fish Speech suffers from severe performance issues with extremely high latency (102 seconds) and an impractical RTF of 31.467.
IndexTTS emerges as the accuracy champion with the lowest Word Error Rate of 0.2418, making it ideal for applications where content fidelity is paramount, such as audiobook production or educational content. The model's controllability features, particularly for Chinese pronunciation, add to its appeal for multilingual applications.
XTTS_v2 provides the most balanced performance profile, leading in both RTF (0.482) and latency (3.36 seconds) while maintaining reasonable accuracy. Its streaming capabilities and overall efficiency make it well-suited for real-time applications like voice assistants or interactive systems.
F5TTS stands out for its resource efficiency, requiring only 2,994 MB of GPU memory while maintaining competitive performance across other metrics. This makes it particularly attractive for edge deployment scenarios or resource-constrained environments.
Use Case Recommendations
For real-time applications requiring immediate response, XTTS_v2 offers the optimal combination of speed and streaming support. Content creators focusing on accuracy should consider IndexTTS for its superior content fidelity. Voice cloning applications benefit from CosyVoice 2.0's excellent speaker similarity, despite its higher computational requirements. Resource-constrained deployments can leverage F5TTS's efficiency without significant quality compromise. Fish Speech, while innovative, requires significant optimization before being practical for most production use cases.
Conclusion
The open-source TTS landscape offers diverse solutions for different requirements, with no single model dominating all performance aspects. The choice between models should align with specific application needs, whether prioritizing speed, accuracy, voice similarity, or resource efficiency. As these models continue to evolve, we expect improvements in streaming implementations, hybrid architectures, and optimization techniques that will further democratize high-quality speech synthesis capabilities.
The rapid development in this space demonstrates the vitality of open-source innovation in AI, providing alternatives that can compete with proprietary solutions while offering the flexibility and control that many applications require.


