csukuangfj commited on
Commit
e0ec0ba
·
1 Parent(s): 08af060

add libriheavy

Browse files
Files changed (1) hide show
  1. model.py +88 -1
model.py CHANGED
@@ -976,6 +976,87 @@ def _get_french_pre_trained_model(
976
  return recognizer
977
 
978
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
979
  @lru_cache(maxsize=10)
980
  def _get_streaming_zipformer_pre_trained_model(
981
  repo_id: str,
@@ -1385,6 +1466,12 @@ english_models = {
1385
  "whisper-base.en": _get_whisper_model,
1386
  "whisper-small.en": _get_whisper_model,
1387
  # "whisper-medium.en": _get_whisper_model,
 
 
 
 
 
 
1388
  "csukuangfj/sherpa-onnx-paraformer-en-2024-03-09": _get_paraformer_pre_trained_model,
1389
  "yfyeung/icefall-asr-gigaspeech-zipformer-2023-10-17": _get_gigaspeech_pre_trained_model_onnx, # noqa
1390
  "wgb14/icefall-asr-gigaspeech-pruned-transducer-stateless2": _get_gigaspeech_pre_trained_model, # noqa
@@ -1448,8 +1535,8 @@ cantonese_models = {
1448
  }
1449
 
1450
  korean_models = {
1451
- "k2-fsa/sherpa-onnx-streaming-zipformer-korean-2024-06-16": _get_streaming_zipformer_pre_trained_model,
1452
  "k2-fsa/sherpa-onnx-zipformer-korean-2024-06-24": _get_offline_pre_trained_model,
 
1453
  }
1454
 
1455
  thai_models = {
 
976
  return recognizer
977
 
978
 
979
+ _get_sherpa_onnx_offline_zipformer_pre_trained_model
980
+
981
+
982
+ @lru_cache(maxsize=10)
983
+ def _get_sherpa_onnx_offline_zipformer_pre_trained_model(
984
+ repo_id: str,
985
+ decoding_method: str,
986
+ num_active_paths: int,
987
+ ) -> sherpa_onnx.OfflineRecognizer:
988
+ assert repo_id in [
989
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-large",
990
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-medium",
991
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-small",
992
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-large-punct-case",
993
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-medium-punct-case",
994
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-small-punct-case",
995
+ ], repo_id
996
+
997
+ if repo_id == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-large":
998
+ epoch = 16
999
+ avg = 3
1000
+ elif repo_id == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-medium":
1001
+ epoch = 60
1002
+ avg = 20
1003
+ elif repo_id == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-small":
1004
+ epoch = 90
1005
+ avg = 20
1006
+ elif (
1007
+ repo_id
1008
+ == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-large-punct-case"
1009
+ ):
1010
+ epoch = 16
1011
+ avg = 2
1012
+ elif (
1013
+ repo_id
1014
+ == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-medium-punct-case"
1015
+ ):
1016
+ epoch = 50
1017
+ avg = 15
1018
+ elif (
1019
+ repo_id
1020
+ == "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-small-punct-case"
1021
+ ):
1022
+ epoch = 88
1023
+ avg = 41
1024
+
1025
+ encoder_model = _get_nn_model_filename(
1026
+ repo_id=repo_id,
1027
+ filename=f"encoder-epoch-{epoch}-avg-{avg}.int8.onnx",
1028
+ subfolder=".",
1029
+ )
1030
+
1031
+ decoder_model = _get_nn_model_filename(
1032
+ repo_id=repo_id,
1033
+ filename=f"decoder-epoch-{epoch}-avg-{avg}.onnx",
1034
+ subfolder=".",
1035
+ )
1036
+
1037
+ joiner_model = _get_nn_model_filename(
1038
+ repo_id=repo_id,
1039
+ filename=f"joiner-epoch-{epoch}-avg-{avg}.int8.onnx",
1040
+ subfolder=".",
1041
+ )
1042
+
1043
+ tokens = _get_token_filename(repo_id=repo_id, subfolder=".")
1044
+
1045
+ recognizer = sherpa_onnx.OfflineRecognizer.from_transducer(
1046
+ tokens=tokens,
1047
+ encoder=encoder_model,
1048
+ decoder=decoder_model,
1049
+ joiner=joiner_model,
1050
+ num_threads=2,
1051
+ sample_rate=16000,
1052
+ feature_dim=80,
1053
+ decoding_method=decoding_method,
1054
+ max_active_paths=num_active_paths,
1055
+ )
1056
+
1057
+ return recognizer
1058
+
1059
+
1060
  @lru_cache(maxsize=10)
1061
  def _get_streaming_zipformer_pre_trained_model(
1062
  repo_id: str,
 
1466
  "whisper-base.en": _get_whisper_model,
1467
  "whisper-small.en": _get_whisper_model,
1468
  # "whisper-medium.en": _get_whisper_model,
1469
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-large": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
1470
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-medium": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
1471
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230926-small": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
1472
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-large-punct-case": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
1473
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-medium-punct-case": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
1474
+ "csukuangfj/sherpa-onnx-zipformer-en-libriheavy-20230830-small-punct-case": _get_sherpa_onnx_offline_zipformer_pre_trained_model,
1475
  "csukuangfj/sherpa-onnx-paraformer-en-2024-03-09": _get_paraformer_pre_trained_model,
1476
  "yfyeung/icefall-asr-gigaspeech-zipformer-2023-10-17": _get_gigaspeech_pre_trained_model_onnx, # noqa
1477
  "wgb14/icefall-asr-gigaspeech-pruned-transducer-stateless2": _get_gigaspeech_pre_trained_model, # noqa
 
1535
  }
1536
 
1537
  korean_models = {
 
1538
  "k2-fsa/sherpa-onnx-zipformer-korean-2024-06-24": _get_offline_pre_trained_model,
1539
+ "k2-fsa/sherpa-onnx-streaming-zipformer-korean-2024-06-16": _get_streaming_zipformer_pre_trained_model,
1540
  }
1541
 
1542
  thai_models = {