Google enters a fast-moving dictation race

Google has quietly released a new iPhone app called Google AI Edge Eloquent, an experimental dictation tool that emphasizes local processing and AI-assisted cleanup of spoken text. The release positions Google more directly against a growing field of voice-to-text products such as Wispr Flow, SuperWhisper and Willow, all of which are trying to turn raw speech recognition into polished writing.

The timing matters because speech interfaces are improving quickly. As automatic speech recognition models become more accurate and smaller models become practical on consumer devices, the market is shifting from plain transcription toward tools that can rewrite spoken language into cleaner, more usable prose. Google’s new app appears designed squarely for that transition.

According to the supplied source text, the app is free on iOS and uses Gemma-based automatic speech recognition models once downloaded to the device. That means core dictation can run locally rather than depending entirely on a cloud connection. In a period when AI products often default to server-side processing, an offline-first design stands out both as a privacy feature and as a reliability feature for users who want speech tools to work in more situations.

More than transcription

The product is not framed as a simple recorder that converts speech into text verbatim. Instead, it aims to bridge the gap between natural speech and ready-to-use writing. The app shows a live transcript while the user is speaking, then performs a second stage of cleanup when the session is paused. The supplied description says it can remove filler words such as “um” and “ah” and smooth over self-corrections to produce cleaner output.

That design choice is important because spoken language and written language are not the same. People pause, restart thoughts and wander mid-sentence in ways that look clumsy on the page even when the intended meaning is clear. AI dictation products are now competing on how well they can infer intent without over-editing the user’s meaning. Google appears to be treating that editorial layer as a core feature rather than an add-on.

The app also includes transformation options labeled “Key points,” “Formal,” “Short” and “Long,” according to the source report. Those controls suggest Google is blending transcription with light text generation. Instead of stopping at accurate capture, the app can reshape the output for different contexts, whether the user wants a summary, a more formal tone or a shorter version.

Local mode and cloud mode point to a hybrid strategy

One of the more notable product details is that cloud processing can be turned off. When cloud mode is enabled, the app uses Gemini models for text cleanup. When it is disabled, the experience remains local-only. That creates a hybrid architecture: on-device models handle the core dictation workflow, while the cloud can be layered in for extra processing when the user wants it.

This is a pragmatic product strategy. Local processing reduces latency, keeps the app working offline and may appeal to users who are cautious about sending audio or drafts to remote servers. Cloud processing, meanwhile, allows more capable text cleanup when a connection is available. Rather than forcing users to choose a strictly local or strictly cloud assistant, Google is testing whether both modes can coexist in one writing tool.

The app can also import certain keywords, names and jargon from a user’s Gmail account if the user opts in. It additionally allows custom words to be added manually. That matters because dictation quality often breaks down on proper nouns, specialist terms and personal vocabulary. Personalized dictionaries can materially improve usefulness, especially for work settings where people routinely speak product names, company terms or technical language that generic models may miss.

Signals beyond the iPhone launch

Although the app is available on iOS now, the source text notes that the App Store description referenced Android integration, including the possibility of using the tool as a default keyboard across text fields and accessing transcription through a floating button. An update cited by the source later removed references to the Android app, while adding that an iOS keyboard is coming soon.

That sequence suggests the release is still in an early, somewhat fluid stage. But it also hints at a larger ambition than a standalone iPhone app. System-wide keyboard access would make the product more strategically important because dictation would no longer be limited to one interface. It could become a layer across messaging, note-taking, email and document workflows.

If Google eventually brings the concept into Android more deeply, it could use platform advantages that smaller rivals cannot easily match. Integration into the default keyboard or broader operating system would give Google a distribution path far beyond a single experimental app. Even if Eloquent remains a test bed, the features being trialed could feed into future transcription and voice features across Google’s mobile ecosystem.

Why this release matters

The most important takeaway is not that Google has launched yet another AI app. It is that the company is testing a product category that sits between speech recognition, editing assistance and personal productivity. That category has become more viable as smaller models improve, and it aligns with a broader industry push to make AI tools feel less like chatbots and more like invisible workflow utilities.

Google’s app also reflects a wider shift in AI product design. Users increasingly want tools that are fast, optionally private and useful in constrained environments. Offline-first software answers those needs directly. If the approach proves successful, it may influence how voice input is built into phones more broadly, especially as users become more comfortable speaking drafts rather than typing them.

For now, Google AI Edge Eloquent looks like an experiment with clear commercial logic. It tests whether users want dictation that does more than transcribe, whether hybrid local-and-cloud processing is compelling, and whether Google can translate advances in speech and language models into a practical everyday tool. In a crowded AI app landscape, that is a more concrete and potentially durable bet than many flashy consumer demos.

This article is based on reporting by TechCrunch. Read the original article.