← Blog·Article·5 min read

How Much Audio Do You Need to Train a Speech Model?

How much audio do you need to train a speech model? Real numbers for ASR, TTS, and conversational AI — plus why hours alone do not tell the story.

The short answer by use case

For fine-tuning a pretrained ASR model on a new domain, plan for 50 to 500 hours of labeled audio. Below 50 you risk overfitting; above 500 the marginal lift drops sharply. For training an ASR model from scratch on a new language, the realistic floor is 1,000 hours and the comfortable target is 10,000.

What drives the answer up or down

Five factors push the required hour count up or down. First, the gap between your current model and your target. A model already strong in your domain needs less data to improve. Second, the quality of the data. High-quality, well-labeled, diverse audio is worth two or three times its hour count compared to noisy, narrow data.

Studio-grade source audio is the bottleneck for production speech AI

Why hours can mislead you

Hour count is a convenient unit but a misleading one. Two corpora of identical size can produce wildly different model accuracy. The corpus that wins is whichever one has the right shape — the right speakers, the right acoustic conditions, the right linguistic distribution — for your deployment.

Real conversation has overlap, repair, and pacing that scripted reads cannot reproduce

How to estimate your specific need

Run the experiment. Take a small slice of candidate data — 10 to 30 hours — and fine-tune your existing baseline. Measure the change in your production metric. Extrapolate. Most fine-tuning curves are roughly logarithmic: each doubling of data produces a smaller absolute improvement.

Realistic budgets for common goals

Adding a new accent to an existing English ASR model: 50 to 150 hours of that accent specifically. Training a new TTS voice: 10 to 20 hours of clean studio audio from one speaker. Adding a new language to a multilingual ASR: 200 to 1,000 hours, depending on the language difficulty.

Per-file provenance is the difference between a defensible dataset and a liability
FAQ

Frequently asked questions

How many hours of audio do I need to train a TTS voice?

Modern neural TTS voices need 5 to 25 hours of clean single-speaker audio for a high-quality result. Smaller amounts work with strong pretrained models but produce less expressive voices.

What is the minimum data to fine-tune Whisper?

Practical Whisper fine-tunes start at 30 to 50 hours for a measurable lift. Below that, you mostly memorize the data. Above 500 hours the marginal improvement drops sharply for most domains.

Do more hours always make a speech model better?

No. Past a point, marginal hours stop moving the needle and useful diversity matters more. A smaller, better-distributed corpus often beats a larger, narrower one.

How much data does a foundation speech model need?

Foundation speech models typically train on 50,000 to 500,000 hours. The leading public systems are trained on hundreds of thousands of hours of mixed-quality audio.

How can I tell when to stop buying more data?

Run a fine-tune experiment with a small slice and extrapolate the lift curve. Buy up to the inflection point where each additional hour produces meaningful lift, then stop.

Looking to license speech data?

Studio-grade conversational audio with aligned transcripts, full speaker metadata, and a documented chain of consent for every file. Get a sample within 48 hours of NDA.

Request a sample →