Research & Development

AI innovation in speech technology and beyond

At Speechmatics, our relentless pursuit of innovation ensures unparalleled accuracy and broad language inclusivity, redefining industry standards.

Expanding accuracy through innovation

Enabling autonomous learning to enhance our understanding

At Speechmatics, self-supervised learning (SSL) serves as a transformative approach in training speech models, harnessing unlabeled data to enhance our speech recognition systems.

This technique allows us to autonomously identify patterns in vast amounts of data, significantly expanding the diversity of speech variations our models can learn from and improving accuracy across multiple languages.

  • FastCompany
  • National Innovation Awards
  • Go:Tech Awards

Tirelessly pushing speech technology forward...

Throughout the years, Speechmatics has remained at the forefront of speech recognition research and innovation. We are consistently pushing the boundaries of what's possible in speech-to-text technology.

2023
Our groundbreaking speech-to-text engine that sets a new benchmark in transcription accuracy. Our next-generation deep learning system trained on millions of hours of audio, is designed to capture spoken words with unparalleled precision, even in noisy or challenging environments.

Our published research

Bias-Augmented Consistency Training Reduces Biased Reasoning in Chain-of-Thought

Bias-Augmented Consistency Training Reduces Biased Reasoning in Chain-of-Thought

James Chua, Edward Rees, Hunar Batra, Samuel R. Bowman, Julian Michael, Ethan Perez, Miles Turpin. March 8, 2024.
While chain-of-thought prompting (CoT) has the potential to improve the explainability of language model reasoning, it can systematically misrepresent the factors influencing models' behavior--for example, rationalizing answers in line with a user's opinion without mentioning this bias. To mitigate this biased reasoning problem, we introduce bias-augmented consistency training (BCT), an unsupervised fine-tuning scheme that trains models to give consistent reasoning across prompts with and without biasing features.Read the paper in full

Debating with More Persuasive LLMs Leads to More Truthful Answers

Debating with More Persuasive LLMs Leads to More Truthful Answers

Akbir Khan, John Hughes, Dan Valentine, Laura Ruis, Kshitij Sachan, Ansh Radhakrishnan, Edward Grefenstette, Samuel R. Bowman, Tim Rocktäschel, Ethan Perez. February 9, 2024.
Common methods for aligning large language models (LLMs) with desired behaviour heavily rely on human-labelled data. However, as models grow increasingly sophisticated, they will surpass human expertise, and the role of human evaluation will evolve into non-experts overseeing experts. In anticipation of this, we ask: can weaker models assess the correctness of stronger models?Read the paper in full

Hierarchical Quantized Autoencoders

Hierarchical Quantized Autoencoders

Will Williams, Sam Ringer, Tom Ash, John Hughes, David MacLeod, Jamie Dougherty. February 19, 2020.
Speechmatics’ paper was submitted and accepted to the most prestigious ML conference – NeurIPs. The paper is about a type of lossy image compression algorithm based on discrete representation learning, leading to a system that can reconstruct images of high-perceptual quality and retain semantically meaningful features despite very high compression rates.Read the paper in full

Texture Bias Of CNNs Limits Few-Shot Classification Performance

Texture Bias Of CNNs Limits Few-Shot Classification Performance

Sam Ringer, Will Williams, Tom Ash, Remi Francis, David MacLeod. October 18, 2019.
Speechmatics published the paper at NeurIPS 2019 presenting in the meta-learning workshop.Read the paper in full

Be part of our mission

We are actively seeking talented individuals to join our collective team of ambitious, problem solvers and throught-leaders, paving the way for inclusion in speech recognition technology.