Cohere has launched Transcribe, a speech recognition model that outperforms OpenAI's Whisper Large v3 and ElevenLabs Scribe v2 on standardized accuracy benchmarks, signaling a major competitive challenge to the incumbents dominating enterprise transcription workflows. The model is available today as free, open-source software and through Cohere's cloud platform, offering developers and businesses a new option for converting audio to text at scale. How Does Cohere Transcribe Compare to Industry Leaders? Cohere Transcribe ranks first on the HuggingFace Open ASR Leaderboard, a standardized benchmark that measures word error rate (WER), the percentage of transcribed words that don't match the original spoken content. Lower scores indicate higher accuracy. The performance gap is substantial. Cohere Transcribe achieved an average word error rate of 5.42%, compared to OpenAI's Whisper Large v3 at 7.44%, representing roughly a 27% relative improvement in accuracy. ElevenLabs Scribe v2 scored 11.86%, placing it further behind. Beyond automated benchmarks, Cohere conducted human evaluations where trained reviewers compared transcripts across real-world audio. In English-language pairwise comparisons, Cohere Transcribe was preferred over Whisper Large v3 in 64% of cases and over ElevenLabs Scribe v2 in 51% of cases. Performance varied by language; Japanese showed a strong advantage at 66-70% against tested rivals, while German and Spanish preference scores hovered around 50%. What Technical Approach Powers This Accuracy Gain? Cohere Transcribe uses a conformer-based encoder-decoder architecture, a design pattern that has become standard in modern speech recognition. A large Conformer encoder extracts acoustic features from audio, while a lightweight Transformer decoder converts those features into text. The company trained the model from scratch rather than fine-tuning an existing model, with deliberate focus on minimizing word error rate under production conditions. At 2 billion parameters, the model strikes a practical balance between accuracy and deployment feasibility. This size is large enough to achieve state-of-the-art performance while remaining practical for real-world graphics processing unit (GPU) deployment or local use on edge hardware. How to Deploy Cohere Transcribe for Your Use Case - Open-Source Download: Access the model directly from HuggingFace and run it locally on your own hardware, including edge devices, with no licensing restrictions under the Apache 2.0 license. - API Access: Experiment for free through Cohere's API dashboard with rate limits, making it easy to test the model's performance on your audio before committing to production deployment. - Managed Cloud Deployment: Use Cohere's Model Vault for production workloads with dedicated infrastructure, no rate limits, and per-hour pricing for organizations requiring enterprise-grade reliability and support. The model supports 14 languages spanning multiple regions. European languages include English, French, German, Italian, Spanish, Portuguese, Greek, Dutch, and Polish. Asia-Pacific coverage includes Chinese, Japanese, Korean, and Vietnamese. Arabic serves the Middle East and North Africa region. Why Does This Matter for Enterprise AI Workflows? Whisper has been the default choice for developers building transcription into products for years, making Cohere's competitive entry significant. The accuracy improvements translate directly to fewer manual corrections, faster processing times, and better user experiences in applications ranging from customer support automation to meeting transcription and accessibility tools. Speed is another critical advantage. The model turns minutes of audio into usable transcripts in seconds, unlocking new possibilities for real-time products and workflows where latency matters. This performance profile makes Cohere Transcribe viable for live customer support interactions, real-time meeting notes, and other time-sensitive applications. Cohere describes this launch as its "zero to one" moment in enterprise speech, positioning it as a starting point rather than a finished product. The company is working toward deeper integration of Transcribe with North, its AI agent orchestration platform, with plans to expand from transcription into broader speech intelligence capabilities such as real-time customer support and speech analytics. For a company that has largely competed on the strength of its large language models and retrieval tools, the move into speech signals a push to cover more of the enterprise AI stack. This represents a direct challenge to incumbents like OpenAI and ElevenLabs on a modality that is increasingly central to automated business workflows, potentially reshaping how organizations choose their transcription infrastructure.