Skip to content

Commit 5bbde3c

Browse files
Open ASR Leaderboard, private datasets (#3374)
* Add files * Update open-asr-leaderboard-private-data.md Co-authored-by: Steven Zheng <58599908+Deep-unlearning@users.noreply.github.com> * Update open-asr-leaderboard-private-data.md Co-authored-by: Steven Zheng <58599908+Deep-unlearning@users.noreply.github.com> --------- Co-authored-by: Steven Zheng <58599908+Deep-unlearning@users.noreply.github.com>
1 parent 90e3475 commit 5bbde3c

3 files changed

Lines changed: 153 additions & 0 deletions

File tree

_blog.yml

Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -5315,3 +5315,11 @@
53155315
- enterprise
53165316
- partnerships
53175317
- hub
5318+
5319+
- local: open-asr-leaderboard-private-data
5320+
date: May 6, 2026
5321+
tags:
5322+
- audio
5323+
- speech
5324+
- leaderboard
5325+
- partnerships
86.3 KB
Loading
Lines changed: 145 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,145 @@
1+
---
2+
title: "Adding Benchmaxxer Repellant to the Open ASR Leaderboard"
3+
thumbnail: /blog/assets/open-asr-leaderboard-private-data/thumbnail.png
4+
authors:
5+
- user: bezzam
6+
- user: Steveeeeeeen
7+
- user: eustlb
8+
- user: SBruccoleriAppen
9+
guest: true
10+
org: AppenAIResearch
11+
- user: jmss-appen
12+
guest: true
13+
org: AppenAIResearch
14+
- user: c-e-ford-appen
15+
guest: true
16+
org: AppenAIResearch
17+
- user: wgb14
18+
guest: true
19+
org: DataoceanAI1
20+
- user: YukaiHuang
21+
guest: true
22+
org: DataoceanAI1
23+
- user: like2026
24+
guest: true
25+
org: DataoceanAI1
26+
- user: logicbean
27+
guest: true
28+
org: DataoceanAI1
29+
- user: ally-lxl
30+
guest: true
31+
org: DataoceanAI1
32+
---
33+
34+
35+
# Adding Benchmaxxer Repellant to the Open ASR Leaderboard
36+
37+
*"When a measure becomes a target, it ceases to be a good measure." (Goodhart’s Law)*
38+
39+
**TLDR**: [Appen Inc.](https://huggingface.co/AppenAIResearch) and [DataoceanAI](https://huggingface.co/DataoceanAI1) have provided high-quality English ASR datasets covering scripted and conversational speech over multiple accents. To prevent potential risks of benchmaxxing or test-set contamination, we will keep these datasets private for a high-quality measure of performance on multiple tasks.
40+
41+
**We’re not updating the average WER at this time**: by default, the leaderboard’s Average WER remains computed on public datasets only. You can optionally include the private datasets using the toggle to see their impact 👀
42+
43+
---
44+
45+
Since its launch in September 2023, the [Open ASR Leaderboard](https://huggingface.co/spaces/hf-audio/open_asr_leaderboard) has been visited over 710K times. We’re blown away by the community’s interest and motivation to keep pushing speech recognition 🗣️
46+
47+
<div align="center">
48+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/leaderboard_stats_may_2026.png" width="512px" alt="thumbnail" />
49+
</div>
50+
51+
Two words sum up the objectives (but also challenges) in maintaining a benchmark like the Open ASR Leaderboard:
52+
53+
1. **Standardization**: models can have different conventions for their usage and outputs, e.g. with/without punctuation and casing. Datasets have the same challenges and can be structured differently. To this end, all test sets have been gathered into a [single dataset](https://huggingface.co/datasets/hf-audio/open-asr-leaderboard) on the Hub for easy access and previewing. Moreover, to standardize model outputs and dataset transcripts, we use a [normalizer](https://github.com/huggingface/open_asr_leaderboard/blob/0009f5fe216d63eea809f9849f4d4534c6ab341e/normalizer/normalizer.py#L528) that (among other things) removes punctuation and casing, and maps to American spelling. It is based on the normalizer of [Whisper](https://github.com/openai/whisper).
54+
55+
2. **Openness**: the [UI code](https://huggingface.co/spaces/hf-audio/open_asr_leaderboard/tree/main) and [evaluation scripts](https://github.com/huggingface/open_asr_leaderboard) are open-sourced. This has helped not only to incorporate new models, but also to improve the quality of the evaluation procedure through community feedback and contributions.
56+
57+
Standardization and openness are essential for meaningful benchmarking, but they also make benchmarks more susceptible to benchmark-specific optimization ("benchmaxxing"), where models improve leaderboard performance without corresponding gains in real-world robustness. As models and use cases evolve, the Open ASR Leaderboard will continue incorporating high-quality datasets and new evaluation settings to better reflect real-world performance and improve robustness against benchmark-specific optimization.
58+
59+
As discussed in our [report](https://arxiv.org/abs/2510.06961), there is no single "catch-all" ASR model: some perform better on American English, others on diverse accents and multilingual settings, while others are optimized for speed or conversational audio. Different applications also prioritize different capabilities, so a model that performs less well on one dimension is not necessarily a worse model overall. The goal of the Open ASR Leaderboard is to capture these nuances and provide a more holistic view of ASR performance.
60+
61+
## New high-quality, private datasets
62+
63+
To this end, we have worked with Appen Inc. and DataoceanAI to curate high-quality datasets for ASR benchmarking. Below is some information on the various splits.
64+
65+
| Dataset | Accent | Duration [h] | Male (%) / Female (%) | Style | Transcription |
66+
| --- | --- | --- | --- | --- | --- |
67+
| Appen Scripted AU | Australian | 1.42 | 49 / 51 | Read | Punctuated, cased. |
68+
| Appen Scripted CA | Canadian | 1.53 | 52 / 48 | Read | Punctuated, cased. |
69+
| Appen Scripted IN | Indian | 1.02 | 49 / 51 | Read | Punctuated, cased. |
70+
| Appen Scripted US | American | 1.45 | 49 / 51 | Read | Punctuated, cased. |
71+
| Appen Conversational IN | Indian | 1.37 | 51 / 49 | Conversational, spontaneous | Punctuated, disfluencies. |
72+
| Appen Conversational US003 | American | 1.64 | 49 / 51 | Conversational, spontaneous | Punctuated, cased, disfluencies. |
73+
| Appen Conversational US004 | American | 1.65 | 49 / 51 | Conversational, spontaneous | Punctuated, disfluencies. |
74+
| DataoceanAI Scripted US | American | 2.43 | 54 / 46 | Read | Punctuated, cased (proper nouns), disfluencies. |
75+
| DataoceanAI Scripted GB | British | 2.43 | 47 / 53 | Read | Punctuated, disfluencies. |
76+
| DataoceanAI Conversational US | American | 8.82 | NA | Conversational, spontaneous | Punctuated, disfluencies. |
77+
| DataoceanAI Conversational GB | British | 5.96 | NA | Conversational, spontaneous | Punctuated, disfluencies. |
78+
79+
Below are sample audio showing the variety of content (scripted, conversational, acronyms, disfluencies, proper nouns).
80+
81+
<audio controls src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/acronym.wav"></audio>
82+
83+
<audio controls src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/proper_noun.wav"></audio>
84+
85+
<audio controls src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/conv.wav"></audio>
86+
87+
<audio controls src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/discluency.wav"></audio>
88+
89+
90+
While private datasets may sound contrary to the spirit of openness, we believe that incorporating such datasets will **increase the trustworthiness** of the Open ASR Leaderboard, as they are less likely to be exploited for benchmaxxing, whether by model developers who explicitly use the public test sets or who try to find training data that closely resembles a particular dataset to boost their score in the macroaverage.
91+
92+
With these datasets, we can also provide targeted metrics to highlight gaps and biases between controlled and often saturated settings (scripted, American accent) and more nuanced conditions (conversational and non-American accents). Below is a screenshot of the new "Private data" tab.
93+
94+
<div align="center">
95+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/private_tab.png" width="1024px" alt="thumbnail" />
96+
</div>
97+
98+
99+
Below is how each column is computed.
100+
- "Average WER" computes a macroaverage of the data provider averages, so that they are weighted equally.
101+
- "Avg Scripted" performs a macroaverage of all scripted datasets.
102+
- "Avg Conversational" performs a macroaverage of all conversational datasets.
103+
- "Avg US" performs a macroaverage of all datasets with American accents.
104+
- "Avg non-US" performs a macroaverage of all datasets with non-American accents.
105+
106+
We intentionally do not provide a score on each split, to avoid model developers from boosting their score with a specific data provider or accent.
107+
108+
## How can I evaluate my model on this data?
109+
110+
Get your model on the Open ASR Leaderboard, and we'll run the evaluation! As before, the process for adding a model to the leaderboard takes place on the Open ASR Leaderboard [GitHub](https://github.com/huggingface/open_asr_leaderboard):
111+
1. Open a pull request, and a [model checklist](https://github.com/huggingface/open_asr_leaderboard/blob/main/.github/PULL_REQUEST_TEMPLATE.md#new-model-checklist) will appear. As before, you should report your results on the public datasets.
112+
2. We will verify the results on the public sets and compute the metrics on the private ones.
113+
3. Confirm the results we’ve obtained.
114+
115+
While you wait for your model to be added to the Open ASR Leaderboard, you can self-report your metrics on the public sets by adding a YAML file like [this](https://huggingface.co/CohereLabs/cohere-transcribe-03-2026/blob/main/.eval_results/open_asr_leaderboard.yaml) to your model card. Your model will then appear on an (unverified) leaderboard that appears on the [dataset page](https://huggingface.co/datasets/hf-audio/open-asr-leaderboard) (see screenshot below). More on this approach to decentralized evaluation can be read [here](https://huggingface.co/blog/community-evals).
116+
117+
<div align="center">
118+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/dataset_leaderboard.png" width="1024px" alt="thumbnail" />
119+
</div>
120+
121+
## Do models trained on the data providers have an advantage?
122+
123+
They could. We’ve asked Appen and DataoceanAI to not provide this data to their clients. But even if they do not provide this exact data, data from a similar distribution could still help the model on the corresponding evaluation set (similar to benchmaxxing by optimizing for a challenging task from the public sets). To this end, having multiple data providers balances out the advantage a model may get from having used data from one of the providers. And we are open to more data providers and eval sets for the "Private data" tab!
124+
125+
Moreover, to ensure that the private sets do not affect the model ranking, we’ve defaulted the Average WER **to not include the Private sets in its macroaverage.**
126+
127+
In the screenshot below, you can see that "Private data" is toggled off. This means that the macroaverage across datasets does not include it.
128+
129+
<div align="center">
130+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/private_off.png" width="1024px" alt="thumbnail" />
131+
</div>
132+
133+
Simply toggle on "Private data" splits to include them in the macroaverage.
134+
135+
<div align="center">
136+
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/open_asr_leaderboard/private_on.png" width="1024px" alt="thumbnail" />
137+
</div>
138+
139+
The "Rank Δ" column shows how the ordering changes relative to the default macroaverage configuration. Including or excluding public datasets also changes the macroaverage, allowing users to tailor the evaluation to the use cases and data distributions most relevant to their application.
140+
141+
## What's next?
142+
143+
We’re excited to hear the community’s feedback on how the new track and dataset toggling features help users identify the model(s) that best fit their application(s). We’re also looking into evaluations that better reflect real-world noisy conditions, and you can expect some news on that 😉
144+
145+
While preparing the private evaluation sets, we took extra care to ensure consistent audio and transcript quality across datasets, including developing tooling to identify challenging cases such as low signal-to-noise conditions or transcript mismatches, since these factors can meaningfully affect WER. More on that in a future post!

0 commit comments

Comments
 (0)