Apr 1, 2026 | Read time min

Speechmatics and Cekura bring real-world STT testing to voice agent pipelines

A new integration gives agent developers a QA layer built for the complexity of the real world.
Speechmatics partners with Cekura - Header (1200 × 900)
Speechmatics
SpeechmaticsEditorial Team

Most voice AI failures do not happen in the demo. They happen in production, with real accents, real background noise, and users who switch mid-sentence between languages. Clean-audio benchmarks rarely surface these issues before deployment. By the time a transcription problem becomes visible, it is already in front of users.

Cambridge-based voice AI company Speechmatics and Cekura, an automated QA platform for conversational AI teams, are today announcing a new integration. The partnership embeds Speechmatics' speech-to-text engine directly into Cekura's testing and production monitoring platform, giving voice agent teams a way to test against the full complexity of production audio at every stage of development and deployment.

For Cekura, the decision to build the integration around Speechmatics came down to performance on the edge cases that matter most:

We were really impressed by Speechmatics' performance on complex medical scribing and seamless mid-sentence language switching. What stood out even more is their commitment to providing independent, unbiased benchmarks. We are excited about what this collaboration means for teams building at the frontier of Voice AI. – Sidhant Kabra, Co-Founder, Cekura.

Cekura supports the complete QA lifecycle, from pre-production simulations and CI/CD pipeline integration through to monitoring of live conversations. Adding Speechmatics to that layer means teams are testing transcription inside a working pipeline, against the conditions it will actually encounter, rather than in isolation.

The practical scope is significant. Teams can assign Speechmatics to specific testing personas to validate agent performance against complex speech patterns and diverse dialects, and simulate multi-speaker audio, noisy environments, and rapid back-and-forth dialogue.

Builders get access to capabilities including advanced speaker intelligence, real-time and recorded media transcription, and a dedicated Medical Model that allows clinical agents to be tested on drug names, dosages, and terminology before any patient interaction occurs, reducing the risk of errors where accuracy carries direct consequences.

The integration also introduces controlled, head-to-head comparisons between STT providers, including Azure, Gemini, and Deepgram, within a consistent testing environment. Teams can evaluate performance against their own audio conditions and user base rather than published benchmarks that may not reflect their production reality.

Most voice agent failures don't happen in the demo, they happen in production, with real accents, real noise, and real complexity. Development teams can now test against those conditions before they go live, with a transcription layer already proven in the world's most demanding environments. – Ricardo Herreros-Symons, Chief Strategy and Revenue Officer, Speechmatics.

To book a demo, visit cekura.ai/expert.

Latest Articles

Carousel slide image
Technical

How to build a microbatching workflow with the Speechmatics API

Build a cleaner path between batch and real time. Learn when micro-batching makes sense, how to chunk audio, submit jobs, stitch JSON, and scale safely with the Speechmatics API.

Speechmatics
SpeechmaticsEditorial Team
Carousel slide image
Product

Alphanumeric speech recognition: why voice assistants mangle SKUs (and how to fix it)

A guide for voice AI engineers, ecommerce platforms and warehouse teams on SKU recognition accuracy voice assistant deployments depend on: why speech recognition systems produce transcription errors on product codes, what to measure when error rates matter, and the fixes that move the needle on order picking, voice ordering and customer-facing voice AI.

Speechmatics
SpeechmaticsEditorial Team
Carousel slide image
Technical

The Adobe story: How we made cloud-grade AI work on your laptop

Behind the build: what it takes to make cloud-grade speech recognition work inside Adobe Premiere, and why Whisper raised the stakes.

Andrew Innes
Andrew InnesChief Architect
Carousel slide image
Company

Adobe and Speechmatics deliver cloud-grade speech recognition on-device for Premiere

Adobe Premiere users can run the most accurate on-device transcription locally; efficient enough for a laptop, powerful enough for professional work.

Speechmatics
SpeechmaticsEditorial Team
Carousel slide image
Use Cases

Best speech-to-text AI guide: APIs, platforms and services compared

Speech-to-text has moved from novelty to enterprise infrastructure. Here's how the leading platforms stack up in 2026 — and how to pick the right one.

Tom Young
Tom YoungDigital Specialist
Speechmatics x Thymia combine medical-grade speech-to-text with clinical-grade voice biomarker intelligence to identify health signals.
News

AI can now understand health signals from 15 seconds of your voice, including fatigue, stress and type 2 diabetes

The joint platform returns transcription and health signals in real time, with no additional hardware required.

Speechmatics
SpeechmaticsEditorial Team