Commit
·
e0ec0ba
1
Parent(s):
08af060
add libriheavy
Browse files
model.py
CHANGED
@@ -976,6 +976,87 @@ def _get_french_pre_trained_model(
|
|
976 |
return recognizer
|
977 |
|
978 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
979 |
@lru_cache(maxsize=10)
|
980 |
def _get_streaming_zipformer_pre_trained_model(
|
981 |
repo_id: str,
|
@@ -1385,6 +1466,12 @@ english_models = {
|
|
1385 |
"whisper-base.en": _get_whisper_model,
|
1386 |
"whisper-small.en": _get_whisper_model,
|
1387 |
# "whisper-medium.en": _get_whisper_model,
|
|
|
|
|
|
|
|
|
|
|
|
|
1388 |
"csukuangfj/sherpa-onnx-paraformer-en-2024-03-09": _get_paraformer_pre_trained_model,
|
1389 |
"yfyeung/icefall-asr-gigaspeech-zipformer-2023-10-17": _get_gigaspeech_pre_trained_model_onnx, # noqa
|
1390 |
"wgb14/icefall-asr-gigaspeech-pruned-transducer-stateless2": _get_gigaspeech_pre_trained_model, # noqa
|
@@ -1448,8 +1535,8 @@ cantonese_models = {
|
|
1448 |
}
|
1449 |
|
1450 |
korean_models = {
|
1451 |
-
"k2-fsa/sherpa-onnx-streaming-zipformer-korean-2024-06-16": _get_streaming_zipformer_pre_trained_model,
|
1452 |
"k2-fsa/sherpa-onnx-zipformer-korean-2024-06-24": _get_offline_pre_trained_model,
|
|
|
1453 |
}
|
1454 |
|
1455 |
thai_models = {
|
|
|
976 |
return recognizer
|
977 |
|
978 |
|
979 |
+
_get_sherpa_onnx_offline_zipformer_pre_trained_model
|
980 |
+
|
981 |
+
|
982 |
+
@lru_cache(maxsize=10)
|
983 |
+
def _get_sherpa_onnx_offline_zipformer_pre_trained_model(
|
984 |
+
repo_id: str,
|
985 |
+
decoding_method: str,
|
986 |
+
num_active_paths: int,
|
987 |
+
) -> sherpa_onnx.OfflineRecognizer:
|
988 |
+
assert repo_id in [
|
989 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-large",
|
990 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-medium",
|
991 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-small",
|
992 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-large-punct-case",
|
993 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-medium-punct-case",
|
994 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-small-punct-case",
|
995 |
+
], repo_id
|
996 |
+
|
997 |
+
if repo_id == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-large":
|
998 |
+
epoch = 16
|
999 |
+
avg = 3
|
1000 |
+
elif repo_id == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-medium":
|
1001 |
+
epoch = 60
|
1002 |
+
avg = 20
|
1003 |
+
elif repo_id == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-small":
|
1004 |
+
epoch = 90
|
1005 |
+
avg = 20
|
1006 |
+
elif (
|
1007 |
+
repo_id
|
1008 |
+
== "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-large-punct-case"
|
1009 |
+
):
|
1010 |
+
epoch = 16
|
1011 |
+
avg = 2
|
1012 |
+
elif (
|
1013 |
+
repo_id
|
1014 |
+
== "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-medium-punct-case"
|
1015 |
+
):
|
1016 |
+
epoch = 50
|
1017 |
+
avg = 15
|
1018 |
+
elif (
|
1019 |
+
repo_id
|
1020 |
+
== "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-small-punct-case"
|
1021 |
+
):
|
1022 |
+
epoch = 88
|
1023 |
+
avg = 41
|
1024 |
+
|
1025 |
+
encoder_model = _get_nn_model_filename(
|
1026 |
+
repo_id=repo_id,
|
1027 |
+
filename=f"encoder-epoch-{epoch}-avg-{avg}.int8.onnx",
|
1028 |
+
subfolder=".",
|
1029 |
+
)
|
1030 |
+
|
1031 |
+
decoder_model = _get_nn_model_filename(
|
1032 |
+
repo_id=repo_id,
|
1033 |
+
filename=f"decoder-epoch-{epoch}-avg-{avg}.onnx",
|
1034 |
+
subfolder=".",
|
1035 |
+
)
|
1036 |
+
|
1037 |
+
joiner_model = _get_nn_model_filename(
|
1038 |
+
repo_id=repo_id,
|
1039 |
+
filename=f"joiner-epoch-{epoch}-avg-{avg}.int8.onnx",
|
1040 |
+
subfolder=".",
|
1041 |
+
)
|
1042 |
+
|
1043 |
+
tokens = _get_token_filename(repo_id=repo_id, subfolder=".")
|
1044 |
+
|
1045 |
+
recognizer = sherpa_onnx.OfflineRecognizer.from_transducer(
|
1046 |
+
tokens=tokens,
|
1047 |
+
encoder=encoder_model,
|
1048 |
+
decoder=decoder_model,
|
1049 |
+
joiner=joiner_model,
|
1050 |
+
num_threads=2,
|
1051 |
+
sample_rate=16000,
|
1052 |
+
feature_dim=80,
|
1053 |
+
decoding_method=decoding_method,
|
1054 |
+
max_active_paths=num_active_paths,
|
1055 |
+
)
|
1056 |
+
|
1057 |
+
return recognizer
|
1058 |
+
|
1059 |
+
|
1060 |
@lru_cache(maxsize=10)
|
1061 |
def _get_streaming_zipformer_pre_trained_model(
|
1062 |
repo_id: str,
|
|
|
1466 |
"whisper-base.en": _get_whisper_model,
|
1467 |
"whisper-small.en": _get_whisper_model,
|
1468 |
# "whisper-medium.en": _get_whisper_model,
|
1469 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-large": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
|
1470 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-medium": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
|
1471 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-small": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
|
1472 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-large-punct-case": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
|
1473 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-medium-punct-case": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
|
1474 |
+
"csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-small-punct-case": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
|
1475 |
"csukuangfj/sherpa-onnx-paraformer-en-2024-03-09": _get_paraformer_pre_trained_model,
|
1476 |
"yfyeung/icefall-asr-gigaspeech-zipformer-2023-10-17": _get_gigaspeech_pre_trained_model_onnx, # noqa
|
1477 |
"wgb14/icefall-asr-gigaspeech-pruned-transducer-stateless2": _get_gigaspeech_pre_trained_model, # noqa
|
|
|
1535 |
}
|
1536 |
|
1537 |
korean_models = {
|
|
|
1538 |
"k2-fsa/sherpa-onnx-zipformer-korean-2024-06-24": _get_offline_pre_trained_model,
|
1539 |
+
"k2-fsa/sherpa-onnx-streaming-zipformer-korean-2024-06-16": _get_streaming_zipformer_pre_trained_model,
|
1540 |
}
|
1541 |
|
1542 |
thai_models = {
|