| References | [1] Hoffmann, Jordan, et al. "Training compute-optimal large language models." arXiv preprint arXiv:2203.15556 (2022). [2] Dao, Tri, et al. "Flashattention: Fast and memory-efficient exact attention with io-awareness." Advances in Neural Information Processing Systems 35 (2022): 16344-16359. [3] Yao, Zhewei, et al. "ZeroQuant: Efficient and affordable post-training quantization for large-scale transformers." Advances in Neural Information Processing Systems 35 (2022): 27168-27183. [4] Fawzi, Alhussein, et al. "Discovering faster matrix multiplication algorithms with reinforcement learning." Nature 610.7930 (2022): 47-53. |
| Authors | Lawrence Atkins & David MacLeod |
| Acknowledgements | Caroline Dockes, Ed Rees, Ellena Reid & Markus Hennerbichler |



![[alt: Concentric circles radiate outward from a central orange icon with a white Speechmatics logo. The background is dark blue, enhancing the orange glow. A thin green line runs horizontally across the lower part of the image.]](/_next/image?url=https%3A%2F%2Fimages.ctfassets.net%2Fyze1aysi0225%2F4jGjYveRLo3sKjzBzMIXXa%2F11e90a40df418658e9c15cb1ecff4e4b%2FBlog_image-wide-carousel.webp&w=3840&q=75)
![[alt: Logo design featuring the text "SPEECHMATICS" alongside a stylized logo for "Cekura," set against a soft green background with subtle curved lines.]](/_next/image?url=https%3A%2F%2Fimages.ctfassets.net%2Fyze1aysi0225%2F39N1Yr95B2jvfd7JKGihq0%2F7b1ca5f8d5db0235b64829dcab16b96a%2FSpeechmatics_partners_with_Cekura-wide-carousel.webp&w=3840&q=75)
![[alt: Smiling man with gray hair sits against a teal background, holding a blank clipboard. He wears a blue sweater and appears relaxed and approachable, suggesting a friendly environment.]](/_next/image?url=https%3A%2F%2Fimages.ctfassets.net%2Fyze1aysi0225%2F2B2UcXrPGOWkeyLII5FGUA%2Ff263f595ae176937bdc93a08b55febcd%2FBlog-header__1_-wide-carousel.webp&w=3840&q=75)