All posts

You hold a button, say a few words, and text appears on your screen almost instantly. The experience feels simple, almost trivial. The machinery behind it is anything but. Voice speech to text is one of the most computationally demanding and intellectually rich problems in applied computer science — and the dramatic improvements in accuracy and speed over the past few years reflect decades of research finally reaching maturity.

Understanding how it works helps you use it better, diagnose accuracy problems, and make smarter choices about which tools to trust with important work.

Step One: Capturing the Audio Signal

The process begins when your microphone converts pressure waves in the air — the physical manifestation of sound — into an electrical signal. That electrical signal is then digitized by an analog-to-digital converter, producing a stream of numerical samples. Each sample represents the amplitude of the sound wave at a specific point in time.

The quality of this initial capture determines a ceiling on everything that follows. A high-quality microphone with low self-noise and a flat frequency response gives the transcription system clean, detailed input to work with. A poor-quality microphone — particularly one with high self-noise, limited frequency range, or poor directional characteristics — introduces artifacts that degrade accuracy at every subsequent stage. No amount of sophisticated modeling downstream can fully compensate for a bad microphone.

The sample rate — how many samples per second the system captures — also matters. Voice speech to text systems typically work with audio sampled at 16,000 Hz, which is sufficient to capture all the frequency information relevant to human speech (roughly 80 Hz to 8,000 Hz) while keeping file sizes and processing requirements manageable.

Step Two: Feature Extraction

Raw audio samples are not directly useful for speech recognition. The next step transforms the raw signal into a compact representation that highlights the acoustically meaningful features of speech — the patterns that distinguish one phoneme from another — while discarding information that is not linguistically relevant.

The most common approach uses a technique called Mel-frequency cepstral coefficients, or MFCCs. This involves dividing the audio into short overlapping windows (typically 25 milliseconds long), computing the frequency spectrum of each window, mapping those frequencies to the Mel scale (which approximates how human hearing perceives pitch), and then applying additional mathematical transforms to separate the spectral envelope from fine structure. The result is a compact, time-varying representation of the audio that captures how the vocal tract is shaping sound at each moment.

Modern neural network approaches often learn their own feature representations directly from raw audio or spectrograms, bypassing the handcrafted MFCC pipeline. These learned representations can capture nuances that fixed feature extraction methods miss, contributing to the accuracy improvements seen in recent years.

Step Three: Acoustic Modeling

Acoustic modeling is where the core recognition happens. A neural network takes the sequence of audio features and produces probability distributions over possible phonemes — the basic sound units of language — at each point in time. Phonemes in English include sounds like "b," "ae" (as in "cat"), "sh," and "n."

Modern acoustic models are trained on hundreds of thousands of hours of labeled speech — audio recordings paired with accurate transcripts. Through training, the model learns the statistical mapping from acoustic features to phoneme probabilities. This training is why modern voice speech to text systems handle the enormous variety of human accents, speaking rates, vocal qualities, and recording conditions reasonably well: they have seen enormous amounts of variation during training.

Step Four: Language Modeling

Phoneme probabilities alone are not sufficient to produce accurate transcription. Many different sequences of phonemes are acoustically plausible for a given utterance, and only a language model can determine which sequence is most likely to be what was actually said.

A language model captures the statistical regularities of language — which words tend to follow which other words, which sequences of words form grammatical sentences, which vocabulary items are common versus rare. When the acoustic model is uncertain between "weather" and "whether," the language model resolves the ambiguity using context: "I need to check the" makes "weather" far more probable than "whether."

Modern speech recognition systems combine the acoustic model and language model into a unified neural architecture — most commonly a transformer-based sequence-to-sequence model. These end-to-end systems learn the mapping from audio directly to text without requiring a separate acoustic model and language model to be trained and combined manually.

Step Five: Decoding

The output of the combined acoustic and language model is not a single transcription — it is a probability distribution over all possible transcriptions. Decoding is the process of finding the most probable transcription given those distributions. Exact search through all possible word sequences is computationally intractable, so practical systems use approximate search algorithms — typically beam search — that explore the most promising candidates efficiently.

Beam search maintains a set of the N most probable partial transcriptions at each step and extends each one with the most probable next word. Candidates that fall below a probability threshold are pruned. This produces a high-quality approximate solution in a fraction of the time required for exact search.

Why Speed and Accuracy Improved So Dramatically

Voice speech to text systems have improved enormously over the past decade, and the improvements accelerated sharply around 2022. Several factors drove this:

What This Means for Your Dictation Experience

When you use Steno to dictate on your Mac or iPhone, you are experiencing the output of all of these steps happening in under a second. The audio from your microphone is captured, feature-extracted, processed through a large neural acoustic-language model, decoded, and delivered as text at your cursor — faster than you can finish the next sentence.

The practical implication is that the accuracy you get today, with good microphone placement and a reasonable speaking environment, is dramatically higher than what was achievable just five years ago. What was once a novelty that required constant correction is now a genuine productivity tool capable of replacing the keyboard for most writing tasks.

Speech recognition is not magic — it is sophisticated probabilistic inference over audio. Understanding the machinery helps you get the most out of it.