File size: 1,854 Bytes
0c6a7c4 0e2118b 0c6a7c4 00d1f8f 0c6a7c4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 |
# wav2vec2-large-xlsr-galician
---
language: gl
datasets:
- OpenSLR 77
- mozilla-foundation common_voice_8_0
metrics:
- wer
tags:
- audio
- automatic-speech-recognition
- speech
- xlsr-fine-tuning-week
license: apache-2.0
model-index:
- name: Galician wav2vec2-large-xlsr-galician
results:
- task:
name: Speech Recognition
type: automatic-speech-recognition
dataset_1:
name: OpenSLR
type: openslr
args: gl
dataset_2:
name: mozilla-foundation
type: common voice
args: gl
metrics:
- name: Test WER
type: wer
value: 7.12
---
# Model
Fine-tuned model for Galician language
Based on the [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) self-supervised model
Fine-tune with audio labelled from [OpenSLR](https://openslr.org/77/) and Mozilla [Common_Voice](https://commonvoice.mozilla.org/gl) (both datasets previously refined)
Check training metrics to see results
# Testing
Make sure that the audio speech input is sampled at 16kHz (mono).
```python
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
model = Wav2Vec2ForCTC.from_pretrained("ifrz/wav2vec2-large-xlsr-galician")
processor = Wav2Vec2Processor.from_pretrained("ifrz/wav2vec2-large-xlsr-galician")
# Reading taken audio clip
import librosa, torch
audio, rate = librosa.load("./gl_test_1.wav", sr = 16000)
# Taking an input value
input_values = processor(audio, sampling_rate=16_000, return_tensors = "pt", padding="longest").input_values
# Storing logits (non-normalized prediction values)
logits = model(input_values).logits
# Storing predicted ids
prediction = torch.argmax(logits, dim = -1)
# Passing the prediction to the tokenzer decode to get the transcription
transcription = processor.batch_decode(prediction)[0]
print(transcription)
``` |