Spaces:
Running
Running
Commit
·
b957022
1
Parent(s):
4b2532b
front end edits
Browse files- __pycache__/constants.cpython-310.pyc +0 -0
- __pycache__/init.cpython-310.pyc +0 -0
- __pycache__/utils_display.cpython-310.pyc +0 -0
- app.py +2 -10
- constants.py +6 -19
- utils_display.py +1 -9
__pycache__/constants.cpython-310.pyc
ADDED
Binary file (8.44 kB). View file
|
|
__pycache__/init.cpython-310.pyc
ADDED
Binary file (3.11 kB). View file
|
|
__pycache__/utils_display.cpython-310.pyc
ADDED
Binary file (2.16 kB). View file
|
|
app.py
CHANGED
@@ -10,16 +10,8 @@ LAST_UPDATED = "Nov 22th 2024"
|
|
10 |
|
11 |
column_names = {
|
12 |
"MODEL": "Model",
|
13 |
-
"Avg.
|
14 |
-
"
|
15 |
-
"AMI WER": "AMI",
|
16 |
-
"Earnings22 WER": "Earnings22",
|
17 |
-
"Gigaspeech WER": "Gigaspeech",
|
18 |
-
"LS Clean WER": "LS Clean",
|
19 |
-
"LS Other WER": "LS Other",
|
20 |
-
"SPGISpeech WER": "SPGISpeech",
|
21 |
-
"Tedlium WER": "Tedlium",
|
22 |
-
"Voxpopuli WER": "Voxpopuli",
|
23 |
}
|
24 |
|
25 |
eval_queue_repo, requested_models, csv_results = load_all_info_from_dataset_hub()
|
|
|
10 |
|
11 |
column_names = {
|
12 |
"MODEL": "Model",
|
13 |
+
"Avg. PER": "Average PER ⬇️",
|
14 |
+
"Avg. PWED": "Avg. PWED ⬆️️",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
}
|
16 |
|
17 |
eval_queue_repo, requested_models, csv_results = load_all_info_from_dataset_hub()
|
constants.py
CHANGED
@@ -11,22 +11,14 @@ EVAL_REQUESTS_PATH = Path("eval_requests")
|
|
11 |
banner_url = "https://huggingface.co/datasets/reach-vb/random-images/resolve/main/asr_leaderboard.png"
|
12 |
BANNER = f'<div style="display: flex; justify-content: space-around;"><img src="{banner_url}" alt="Banner" style="width: 40vw; min-width: 300px; max-width: 600px;"> </div>'
|
13 |
|
14 |
-
TITLE = "<html> <head> <style> h1 {text-align: center;} </style> </head> <body> <h1> 🤗
|
15 |
|
16 |
-
INTRODUCTION_TEXT = "📐 The 🤗
|
17 |
on the Hugging Face Hub. \
|
18 |
-
\nWe report the Average [
|
19 |
\nIf you want results for a model that is not listed here, you can submit a request for it to be included ✉️✨. \
|
20 |
\nThe leaderboard currently focuses on English speech recognition, and will be expanded to multilingual evaluation in later versions."
|
21 |
|
22 |
-
CITATION_TEXT = """@misc{open-asr-leaderboard,
|
23 |
-
title = {Open Automatic Speech Recognition Leaderboard},
|
24 |
-
author = {Srivastav, Vaibhav and Majumdar, Somshubra and Koluguri, Nithin and Moumen, Adel and Gandhi, Sanchit and others},
|
25 |
-
year = 2023,
|
26 |
-
publisher = {Hugging Face},
|
27 |
-
howpublished = "\\url{https://huggingface.co/spaces/hf-audio/open_asr_leaderboard}"
|
28 |
-
}
|
29 |
-
"""
|
30 |
|
31 |
METRICS_TAB_TEXT = """
|
32 |
Here you will find details about the speech recognition metrics and datasets reported in our leaderboard.
|
@@ -101,16 +93,11 @@ a model is likely to perform on downstream ASR compared to evaluating it on one
|
|
101 |
The ESB score is calculated as a macro-average of the WER scores across the ESB datasets. The models in the leaderboard
|
102 |
are ranked based on their average WER scores, from lowest to highest.
|
103 |
|
|
|
|
|
104 |
| Dataset | Domain | Speaking Style | Train (h) | Dev (h) | Test (h) | Transcriptions | License |
|
105 |
|-----------------------------------------------------------------------------------------|-----------------------------|-----------------------|-----------|---------|----------|--------------------|-----------------|
|
106 |
-
| [
|
107 |
-
| [VoxPopuli](https://huggingface.co/datasets/facebook/voxpopuli) | European Parliament | Oratory | 523 | 5 | 5 | Punctuated | CC0 |
|
108 |
-
| [TED-LIUM](https://huggingface.co/datasets/LIUM/tedlium) | TED talks | Oratory | 454 | 2 | 3 | Normalised | CC-BY-NC-ND 3.0 |
|
109 |
-
| [GigaSpeech](https://huggingface.co/datasets/speechcolab/gigaspeech) | Audiobook, podcast, YouTube | Narrated, spontaneous | 2500 | 12 | 40 | Punctuated | apache-2.0 |
|
110 |
-
| [SPGISpeech](https://huggingface.co/datasets/kensho/spgispeech) | Financial meetings | Oratory, spontaneous | 4900 | 100 | 100 | Punctuated & Cased | User Agreement |
|
111 |
-
| [Earnings-22](https://huggingface.co/datasets/revdotcom/earnings22) | Financial meetings | Oratory, spontaneous | 105 | 5 | 5 | Punctuated & Cased | CC-BY-SA-4.0 |
|
112 |
-
| [AMI](https://huggingface.co/datasets/edinburghcstr/ami) | Meetings | Spontaneous | 78 | 9 | 9 | Punctuated & Cased | CC-BY-4.0 |
|
113 |
-
|
114 |
For more details on the individual datasets and how models are evaluated to give the ESB score, refer to the [ESB paper](https://arxiv.org/abs/2210.13352).
|
115 |
"""
|
116 |
|
|
|
11 |
banner_url = "https://huggingface.co/datasets/reach-vb/random-images/resolve/main/asr_leaderboard.png"
|
12 |
BANNER = f'<div style="display: flex; justify-content: space-around;"><img src="{banner_url}" alt="Banner" style="width: 40vw; min-width: 300px; max-width: 600px;"> </div>'
|
13 |
|
14 |
+
TITLE = "<html> <head> <style> h1 {text-align: center;} </style> </head> <body> <h1> 🤗 IPA Transcription Leaderboard </b> </body> </html>"
|
15 |
|
16 |
+
INTRODUCTION_TEXT = "📐 The 🤗 IPA transcription Leaderboard ranks and evaluates speech recognition models \
|
17 |
on the Hugging Face Hub. \
|
18 |
+
\nWe report the Average [PER](https://huggingface.co/spaces/evaluate-metric/wer) (⬇️ lower the better) and [RTFx](https://github.com/NVIDIA/DeepLearningExamples/blob/master/Kaldi/SpeechRecognition/README.md#metrics) (⬆️ higher the better). Models are ranked based on their Average WER, from lowest to highest. Check the 📈 Metrics tab to understand how the models are evaluated. \
|
19 |
\nIf you want results for a model that is not listed here, you can submit a request for it to be included ✉️✨. \
|
20 |
\nThe leaderboard currently focuses on English speech recognition, and will be expanded to multilingual evaluation in later versions."
|
21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
|
23 |
METRICS_TAB_TEXT = """
|
24 |
Here you will find details about the speech recognition metrics and datasets reported in our leaderboard.
|
|
|
93 |
The ESB score is calculated as a macro-average of the WER scores across the ESB datasets. The models in the leaderboard
|
94 |
are ranked based on their average WER scores, from lowest to highest.
|
95 |
|
96 |
+
We are currently working to add and curate more datasets. Right now, models will be evaluated just on the TIMIT test dataset for phoneme transcription.
|
97 |
+
|
98 |
| Dataset | Domain | Speaking Style | Train (h) | Dev (h) | Test (h) | Transcriptions | License |
|
99 |
|-----------------------------------------------------------------------------------------|-----------------------------|-----------------------|-----------|---------|----------|--------------------|-----------------|
|
100 |
+
| [TIMIT Dataset](https://www.kaggle.com/datasets/mfekadu/darpa-timit-acousticphonetic-continuous-speech) | Audiobook | Narrated | 960 | 11 | 11 | Normalised | CC-BY-4.0 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
101 |
For more details on the individual datasets and how models are evaluated to give the ESB score, refer to the [ESB paper](https://arxiv.org/abs/2210.13352).
|
102 |
"""
|
103 |
|
utils_display.py
CHANGED
@@ -14,15 +14,7 @@ def fields(raw_class):
|
|
14 |
class AutoEvalColumn: # Auto evals column
|
15 |
model = ColumnContent("Model", "markdown")
|
16 |
avg_wer = ColumnContent("Average WER ⬇️", "number")
|
17 |
-
|
18 |
-
ami_wer = ColumnContent("AMI", "number")
|
19 |
-
e22_wer = ColumnContent("Earnings22", "number")
|
20 |
-
gs_wer = ColumnContent("Gigaspeech", "number")
|
21 |
-
lsc_wer = ColumnContent("LS Clean", "number")
|
22 |
-
lso_wer = ColumnContent("LS Other", "number")
|
23 |
-
ss_wer = ColumnContent("SPGISpeech", "number")
|
24 |
-
tl_wer = ColumnContent("Tedlium", "number")
|
25 |
-
vp_wer = ColumnContent("Voxpopuli", "number")
|
26 |
|
27 |
|
28 |
def make_clickable_model(model_name):
|
|
|
14 |
class AutoEvalColumn: # Auto evals column
|
15 |
model = ColumnContent("Model", "markdown")
|
16 |
avg_wer = ColumnContent("Average WER ⬇️", "number")
|
17 |
+
avg_wped = ColumnContent("Average PWED ⬇️", "number")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
|
19 |
|
20 |
def make_clickable_model(model_name):
|