arunasrivastava commited on
Commit
b957022
·
1 Parent(s): 4b2532b

front end edits

Browse files
__pycache__/constants.cpython-310.pyc ADDED
Binary file (8.44 kB). View file
 
__pycache__/init.cpython-310.pyc ADDED
Binary file (3.11 kB). View file
 
__pycache__/utils_display.cpython-310.pyc ADDED
Binary file (2.16 kB). View file
 
app.py CHANGED
@@ -10,16 +10,8 @@ LAST_UPDATED = "Nov 22th 2024"
10
 
11
  column_names = {
12
  "MODEL": "Model",
13
- "Avg. WER": "Average WER ⬇️",
14
- "RTFx": "RTFx ⬆️️",
15
- "AMI WER": "AMI",
16
- "Earnings22 WER": "Earnings22",
17
- "Gigaspeech WER": "Gigaspeech",
18
- "LS Clean WER": "LS Clean",
19
- "LS Other WER": "LS Other",
20
- "SPGISpeech WER": "SPGISpeech",
21
- "Tedlium WER": "Tedlium",
22
- "Voxpopuli WER": "Voxpopuli",
23
  }
24
 
25
  eval_queue_repo, requested_models, csv_results = load_all_info_from_dataset_hub()
 
10
 
11
  column_names = {
12
  "MODEL": "Model",
13
+ "Avg. PER": "Average PER ⬇️",
14
+ "Avg. PWED": "Avg. PWED ⬆️️",
 
 
 
 
 
 
 
 
15
  }
16
 
17
  eval_queue_repo, requested_models, csv_results = load_all_info_from_dataset_hub()
constants.py CHANGED
@@ -11,22 +11,14 @@ EVAL_REQUESTS_PATH = Path("eval_requests")
11
  banner_url = "https://huggingface.co/datasets/reach-vb/random-images/resolve/main/asr_leaderboard.png"
12
  BANNER = f'<div style="display: flex; justify-content: space-around;"><img src="{banner_url}" alt="Banner" style="width: 40vw; min-width: 300px; max-width: 600px;"> </div>'
13
 
14
- TITLE = "<html> <head> <style> h1 {text-align: center;} </style> </head> <body> <h1> 🤗 Open Automatic Speech Recognition Leaderboard </b> </body> </html>"
15
 
16
- INTRODUCTION_TEXT = "📐 The 🤗 Open ASR Leaderboard ranks and evaluates speech recognition models \
17
  on the Hugging Face Hub. \
18
- \nWe report the Average [WER](https://huggingface.co/spaces/evaluate-metric/wer) (⬇️ lower the better) and [RTFx](https://github.com/NVIDIA/DeepLearningExamples/blob/master/Kaldi/SpeechRecognition/README.md#metrics) (⬆️ higher the better). Models are ranked based on their Average WER, from lowest to highest. Check the 📈 Metrics tab to understand how the models are evaluated. \
19
  \nIf you want results for a model that is not listed here, you can submit a request for it to be included ✉️✨. \
20
  \nThe leaderboard currently focuses on English speech recognition, and will be expanded to multilingual evaluation in later versions."
21
 
22
- CITATION_TEXT = """@misc{open-asr-leaderboard,
23
- title = {Open Automatic Speech Recognition Leaderboard},
24
- author = {Srivastav, Vaibhav and Majumdar, Somshubra and Koluguri, Nithin and Moumen, Adel and Gandhi, Sanchit and others},
25
- year = 2023,
26
- publisher = {Hugging Face},
27
- howpublished = "\\url{https://huggingface.co/spaces/hf-audio/open_asr_leaderboard}"
28
- }
29
- """
30
 
31
  METRICS_TAB_TEXT = """
32
  Here you will find details about the speech recognition metrics and datasets reported in our leaderboard.
@@ -101,16 +93,11 @@ a model is likely to perform on downstream ASR compared to evaluating it on one
101
  The ESB score is calculated as a macro-average of the WER scores across the ESB datasets. The models in the leaderboard
102
  are ranked based on their average WER scores, from lowest to highest.
103
 
 
 
104
  | Dataset | Domain | Speaking Style | Train (h) | Dev (h) | Test (h) | Transcriptions | License |
105
  |-----------------------------------------------------------------------------------------|-----------------------------|-----------------------|-----------|---------|----------|--------------------|-----------------|
106
- | [LibriSpeech](https://huggingface.co/datasets/librispeech_asr) | Audiobook | Narrated | 960 | 11 | 11 | Normalised | CC-BY-4.0 |
107
- | [VoxPopuli](https://huggingface.co/datasets/facebook/voxpopuli) | European Parliament | Oratory | 523 | 5 | 5 | Punctuated | CC0 |
108
- | [TED-LIUM](https://huggingface.co/datasets/LIUM/tedlium) | TED talks | Oratory | 454 | 2 | 3 | Normalised | CC-BY-NC-ND 3.0 |
109
- | [GigaSpeech](https://huggingface.co/datasets/speechcolab/gigaspeech) | Audiobook, podcast, YouTube | Narrated, spontaneous | 2500 | 12 | 40 | Punctuated | apache-2.0 |
110
- | [SPGISpeech](https://huggingface.co/datasets/kensho/spgispeech) | Financial meetings | Oratory, spontaneous | 4900 | 100 | 100 | Punctuated & Cased | User Agreement |
111
- | [Earnings-22](https://huggingface.co/datasets/revdotcom/earnings22) | Financial meetings | Oratory, spontaneous | 105 | 5 | 5 | Punctuated & Cased | CC-BY-SA-4.0 |
112
- | [AMI](https://huggingface.co/datasets/edinburghcstr/ami) | Meetings | Spontaneous | 78 | 9 | 9 | Punctuated & Cased | CC-BY-4.0 |
113
-
114
  For more details on the individual datasets and how models are evaluated to give the ESB score, refer to the [ESB paper](https://arxiv.org/abs/2210.13352).
115
  """
116
 
 
11
  banner_url = "https://huggingface.co/datasets/reach-vb/random-images/resolve/main/asr_leaderboard.png"
12
  BANNER = f'<div style="display: flex; justify-content: space-around;"><img src="{banner_url}" alt="Banner" style="width: 40vw; min-width: 300px; max-width: 600px;"> </div>'
13
 
14
+ TITLE = "<html> <head> <style> h1 {text-align: center;} </style> </head> <body> <h1> 🤗 IPA Transcription Leaderboard </b> </body> </html>"
15
 
16
+ INTRODUCTION_TEXT = "📐 The 🤗 IPA transcription Leaderboard ranks and evaluates speech recognition models \
17
  on the Hugging Face Hub. \
18
+ \nWe report the Average [PER](https://huggingface.co/spaces/evaluate-metric/wer) (⬇️ lower the better) and [RTFx](https://github.com/NVIDIA/DeepLearningExamples/blob/master/Kaldi/SpeechRecognition/README.md#metrics) (⬆️ higher the better). Models are ranked based on their Average WER, from lowest to highest. Check the 📈 Metrics tab to understand how the models are evaluated. \
19
  \nIf you want results for a model that is not listed here, you can submit a request for it to be included ✉️✨. \
20
  \nThe leaderboard currently focuses on English speech recognition, and will be expanded to multilingual evaluation in later versions."
21
 
 
 
 
 
 
 
 
 
22
 
23
  METRICS_TAB_TEXT = """
24
  Here you will find details about the speech recognition metrics and datasets reported in our leaderboard.
 
93
  The ESB score is calculated as a macro-average of the WER scores across the ESB datasets. The models in the leaderboard
94
  are ranked based on their average WER scores, from lowest to highest.
95
 
96
+ We are currently working to add and curate more datasets. Right now, models will be evaluated just on the TIMIT test dataset for phoneme transcription.
97
+
98
  | Dataset | Domain | Speaking Style | Train (h) | Dev (h) | Test (h) | Transcriptions | License |
99
  |-----------------------------------------------------------------------------------------|-----------------------------|-----------------------|-----------|---------|----------|--------------------|-----------------|
100
+ | [TIMIT Dataset](https://www.kaggle.com/datasets/mfekadu/darpa-timit-acousticphonetic-continuous-speech) | Audiobook | Narrated | 960 | 11 | 11 | Normalised | CC-BY-4.0 |
 
 
 
 
 
 
 
101
  For more details on the individual datasets and how models are evaluated to give the ESB score, refer to the [ESB paper](https://arxiv.org/abs/2210.13352).
102
  """
103
 
utils_display.py CHANGED
@@ -14,15 +14,7 @@ def fields(raw_class):
14
  class AutoEvalColumn: # Auto evals column
15
  model = ColumnContent("Model", "markdown")
16
  avg_wer = ColumnContent("Average WER ⬇️", "number")
17
- rtf = ColumnContent("RTFx ⬆️️", "number")
18
- ami_wer = ColumnContent("AMI", "number")
19
- e22_wer = ColumnContent("Earnings22", "number")
20
- gs_wer = ColumnContent("Gigaspeech", "number")
21
- lsc_wer = ColumnContent("LS Clean", "number")
22
- lso_wer = ColumnContent("LS Other", "number")
23
- ss_wer = ColumnContent("SPGISpeech", "number")
24
- tl_wer = ColumnContent("Tedlium", "number")
25
- vp_wer = ColumnContent("Voxpopuli", "number")
26
 
27
 
28
  def make_clickable_model(model_name):
 
14
  class AutoEvalColumn: # Auto evals column
15
  model = ColumnContent("Model", "markdown")
16
  avg_wer = ColumnContent("Average WER ⬇️", "number")
17
+ avg_wped = ColumnContent("Average PWED ⬇️", "number")
 
 
 
 
 
 
 
 
18
 
19
 
20
  def make_clickable_model(model_name):