All posts

Speech to speech technology is one of the most intriguing areas in modern computing. At its broadest definition, it describes any pipeline that begins with spoken input and produces some form of spoken or written output — from real-time translation that converts your English into spoken Mandarin, to dictation systems that convert your voice into written text at the cursor. Understanding the different meanings of "speech to speech" and where voice-to-text tools fit in this landscape helps clarify what you actually need.

In 2026, speech-related technologies have advanced to the point where the distinctions between these systems — once very sharp — have begun to blur. Dictation apps now apply intelligent rewriting that resembles translation. Translation apps now handle the full cycle from source speech to target speech without requiring text as an intermediate step. The voice input stack has become a sophisticated pipeline.

Three Distinct Meanings of "Speech to Speech"

Voice Translation (Speak in One Language, Hear Another)

The most dramatic form of speech to speech is real-time spoken language translation. You speak in English, and within a second or two, a synthesized voice or earpiece delivers the same content in another language. This technology is used for international business calls, travel communication, and accessibility services for the deaf-blind community via tactile interfaces.

Real-time voice translation pipelines are technically demanding. They require accurate speech recognition in the source language, high-quality machine translation that preserves nuance and registers formality correctly, and natural-sounding speech synthesis in the target language — all completed in under a second to maintain conversational flow. The technology has matured dramatically and is genuinely useful for many real-world situations, though edge cases (accents, idioms, technical vocabulary) still trip up the best systems.

Voice Dictation (Speak, Get Text)

This is what most people mean when they think about voice input for everyday productivity. You speak, the system recognizes your words, and text appears in whatever application you are using. The output is written text, not synthesized speech, but the input is always spoken. This category — voice to text dictation — is the most practically impactful speech technology for day-to-day work.

Voice Cloning and AI Voice Generation (Text or Voice In, Custom Voice Out)

A newer category uses speech as training data to create personalized voice models that can then synthesize speech in your voice from any text. This is used for accessibility (people who lose the ability to speak can use a voice clone trained before the loss), content creation (narrating written content in a natural voice without recording sessions), and communication personalization.

Why Voice Dictation Is the Most Impactful for Everyday Users

Of the three categories above, voice dictation — the speech to text pipeline — provides the most immediate, daily-use productivity benefit for the majority of knowledge workers. The math is simple: people speak at 120–150 words per minute and type at 40–70 words per minute. Converting from typing to speaking as the primary text input mode can more than double the speed at which you produce written content.

This productivity gain applies to everything you write at a keyboard: emails, reports, code comments, Slack messages, documentation, social posts, notes, and any other text output that currently flows through your fingers. The constraint is no longer how fast you can physically type — it becomes how clearly you can articulate your thoughts, which is a cognitive skill most people have developed naturally over a lifetime of conversation.

The Role of Language Models in Modern Voice Input

Modern voice input systems are not purely acoustic-to-text converters. They incorporate language models that understand the structure and semantics of language to produce output that reads well, not just output that is acoustically accurate. This intelligence layer is what handles punctuation inference, context-appropriate word choices, and the filtering of speech artifacts that should not appear in written text.

In practical terms, this means that the best modern voice dictation systems produce text that reads like it was written, not transcribed. The spoken-to-written translation happens automatically and invisibly, which is the quality marker that separates professional-grade dictation tools from consumer-grade ones.

How Steno Implements the Speech to Text Pipeline

Steno's architecture is a carefully designed speech-to-text pipeline optimized for professional use on Mac and iPhone. The pipeline operates in four stages.

Stage 1: Audio Capture

When you hold the hotkey on Mac or the dictation button on iPhone, Steno begins capturing audio through the active microphone. Voice isolation processing is applied to reduce background noise and focus capture on your voice specifically, improving the signal quality sent to the recognition engine.

Stage 2: Speech Recognition

The captured audio is processed by a high-accuracy speech recognition model to produce a raw word sequence. This stage handles phoneme recognition, word boundary detection, and acoustic ambiguity resolution using contextual language model priors.

Stage 3: Intelligent Post-Processing

The raw transcription is passed through Steno's language intelligence layer, which applies punctuation, handles capitalization, removes filler words, and adjusts for the domain context established by your voice profile. If Smart Rewrite is enabled, a more aggressive rewrite pass produces polished written prose from your spoken input.

Stage 4: Text Insertion

The final processed text is inserted at your current cursor position in whatever application is active. On Mac, this uses the accessibility API to insert text anywhere. On iPhone, this uses the keyboard extension API. The insertion is instant — you see the result immediately when you release the dictation key.

Speech to Speech Translation on Mac and iPhone

While Steno focuses on voice dictation (speech to text) rather than spoken language translation (speech to speech translation), the two technologies increasingly appear together in professional workflows. A common scenario: a multilingual professional uses voice dictation to capture notes in their native language, then uses a translation layer to convert those notes for international colleagues.

This combined workflow — speak, transcribe, translate — is faster than typing content in any language, let alone a second language. Voice dictation handles the first leg (getting ideas from brain to text quickly), and translation handles the second leg (converting that text for international use). The result is multilingual content production at speeds that would have been impossible without voice input.

Speech to speech is not one technology but a family of technologies. The one that matters most for your daily work is voice dictation — converting your fastest communication channel (speech) into your most versatile output format (text).

If you are on Mac and want to experience the voice dictation side of the speech technology spectrum, download Steno at stenofast.com and try the hold-to-speak workflow for a week.