Aug 31, 2022 | Read time 2 min

Speechmatics launches Language Identification, allowing users to automatically determine the predominant language in a media file

Speechmatics launches Language Identification, allowing users to automatically determine the predominant language in a media file
Speechmatics launches Language Identification
Speechmatics
SpeechmaticsEditorial team

Latest addition to Speechmatics engine saves time on manually reviewing files and is applicable to a wide variety of use cases

Speechmatics, the leading autonomous speech recognition technology scaleup has now added Language Identification (Language ID) to its industry-leading speech-to-text engine. This latest addition allows customers to automatically identify the predominant language spoken in any media file. Customers will save time and effort on manually reviewing files, safe in the knowledge that they will be provided with an accurate transcription of any media file.

Language ID drives efficiency by removing the manual step of selecting which language pack should be used when the language is not explicitly stated on the file. Often requested, it not only helps users identify unknown languages, but also adds useful metadata about the language of the spoken audio. Media and broadcast organizations have extensive archives of audio, the content of which is often unknown. Instead of manually listening to hours of speech – and relying on human interpretation to label it – Speechmatics Language ID confirms the language pre-transcription. For contact centers, being able to identify the predominant language spoken (especially when callers switch languages) is a huge benefit to those conducting call analysis.

Speechmatics has built the most accurate and inclusive speech-to-text engine available. Historically, training data had to be manually tagged, classified or ‘labelled’. This has resulted in engines that have been trained on narrow datasets, which fail to represent the diversity of voices that use them. In contrast, Speechmatics’ speech-to-text engine is trained through exposure to hundreds of thousands of individual voices using millions of hours of unlabelled, more representative voice data. Speechmatics has applied this technique to identifying predominant spoken languages on a diverse set of audio data.

Commenting on this rollout of Language ID, CEO Katy Wigdahl said, “Up until now, identifying languages without human intervention has been costly and time-consuming for users of speech-to-text. However, with our new Language ID, this will be a thing of the past and allow customers to swiftly identify and transcribe media files - with less hassle and more efficiency. We can’t wait for our customers to use this Language ID and see it deliver accurate and valuable results.’’

This latest update can be used with pre-recorded media files, works with up to 12 languages and adds a confidence score to show the certainty of the predominant language. Supported languages are English, German, Spanish, French, Hindi, Italian, Japanese, Korean, Mandarin, Dutch, Portuguese, and Russian.

Latest Articles

[alt: Bilingual medical model featuring terms related to various health conditions and medications in Arabic and English. Key terms include "Chronic kidney disease," "Heart attack," "Diabetes," and "Insulin," among others, displayed in an organized layout.]
Product

Speechmatics achieves a world first in bilingual Voice AI with new Arabic–English model

Sets a new accuracy bar for real-world code-switching: 35% fewer errors than the closest competitor.

Speechmatics
SpeechmaticsEditorial Team
[alt: Illuminated ancient mud-brick structures stand against a dusk sky, showcasing architectural details and textures. Palm trees are in the foreground, adding to the setting's ambiance. Visually captures a historic site in twilight.]
Product

Your voice agent speaks perfect Arabic. That's the problem.

Most voice AI models are trained on formal Arabic, but real conversations across the Middle East mix dialects and English in ways those systems aren’t built to handle.

Yahia Abaza
Yahia AbazaSenior Product Manger
new blog image header
Technical

How Nvidia Dominates the HuggingFace Leaderboards in This Key Metric

A technical deep-dive into Token Duration Transducers (TDT) — the frame-skipping architecture behind Nvidia's Parakeet models. Covers inference mechanics, training with forward-backward algorithm, and how TDT achieves up to 2.82x faster decoding than standard RNN-T.

Oliver Parish
Oliver Parish Machine Learning Engineer
[alt: Healthcare professionals in scrubs and lab coats walk briskly down a hospital corridor. A nurse uses a tablet while others carry patient charts and attend to a gurney. The setting conveys a busy, clinical environment focused on patient care.]
Use Cases

Why AI-native EHR platforms will treat speech as core infrastructure in 2026

As clinical workflows become automated and AI-driven, real-time speech is shifting from a transcription feature to the foundational intelligence layer inside modern EHR systems.

Vamsi Edara
Vamsi EdaraFounder and CEO, Edvak EHR
[alt: Logos of Speechmatics and Edvak are displayed side by side, interconnected by a stylized x symbol. The background features soft, wavy lines in light blue, creating a modern and tech-focused aesthetic.]
Company

One word changes everything: Speechmatics and Edvak EHR partner to make voice AI safe for clinical automation at scale

Turning real-time clinical speech into trusted, EHR-native automation.

Speechmatics
SpeechmaticsEditorial Team
[alt: Concentric circles radiate outward from a central orange icon with a white Speechmatics logo. The background is dark blue, enhancing the orange glow. A thin green line runs horizontally across the lower part of the image.]
Technical

Speed you can trust: The STT metrics that matter for voice agents

What “fast” actually means for voice agents — and why Pipecat’s TTFS + semantic accuracy is the clearest benchmark we’ve seen.

Archie McMullan
Archie McMullanSpeechmatics Graduate