multiple decoding
Browse files
README.md
CHANGED
@@ -63,9 +63,17 @@ Please notice that we encourage you to read our tutorials and learn more about
|
|
63 |
from speechbrain.inference.ASR import EncoderDecoderASR
|
64 |
|
65 |
asr_model = EncoderDecoderASR.from_hparams(source="speechbrain/asr-conformer-largescaleasr", savedir="pretrained_models/asr-conformer-largescaleasr")
|
|
|
|
|
66 |
asr_model.transcribe_file("speechbrain/asr-conformer-largescaleasr/example.wav")
|
67 |
|
|
|
|
|
|
|
|
|
|
|
68 |
```
|
|
|
69 |
### Inference on GPU
|
70 |
To perform inference on the GPU, add `run_opts={"device":"cuda"}` when calling the `from_hparams` method.
|
71 |
|
|
|
63 |
from speechbrain.inference.ASR import EncoderDecoderASR
|
64 |
|
65 |
asr_model = EncoderDecoderASR.from_hparams(source="speechbrain/asr-conformer-largescaleasr", savedir="pretrained_models/asr-conformer-largescaleasr")
|
66 |
+
|
67 |
+
# For a full decoding with a large beam size (can be slow):
|
68 |
asr_model.transcribe_file("speechbrain/asr-conformer-largescaleasr/example.wav")
|
69 |
|
70 |
+
# For smaller beam size:
|
71 |
+
asr_model.transcribe_file("speechbrain/asr-conformer-largescaleasr/example.wav", overrides={"test_beam_size":"10"})
|
72 |
+
|
73 |
+
# For even faster decoding
|
74 |
+
asr_model.transcribe_file("speechbrain/asr-conformer-largescaleasr/example.wav", overrides={"test_beam_size":"10", "ctc_weight_decode":0.0})
|
75 |
```
|
76 |
+
|
77 |
### Inference on GPU
|
78 |
To perform inference on the GPU, add `run_opts={"device":"cuda"}` when calling the `from_hparams` method.
|
79 |
|