Commit
·
4c746d7
1
Parent(s):
edc48b2
add a new model
Browse files- model.py +25 -5
- requirements.txt +2 -2
model.py
CHANGED
|
@@ -178,6 +178,10 @@ def get_pretrained_model(
|
|
| 178 |
return more_than_1600_languages_models[repo_id](
|
| 179 |
repo_id, decoding_method=decoding_method, num_active_paths=num_active_paths
|
| 180 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 181 |
elif repo_id in twenty_five_languages_models:
|
| 182 |
return twenty_five_languages_models[repo_id](
|
| 183 |
repo_id, decoding_method=decoding_method, num_active_paths=num_active_paths
|
|
@@ -2031,15 +2035,24 @@ def _get_sense_voice_pre_trained_model(
|
|
| 2031 |
num_active_paths: int,
|
| 2032 |
) -> sherpa_onnx.OfflineRecognizer:
|
| 2033 |
assert repo_id in [
|
|
|
|
|
|
|
| 2034 |
"csukuangfj/sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17",
|
| 2035 |
"csukuangfj/sherpa-onnx-sense-voice-zh-en-ja-ko-yue-int8-2025-09-09",
|
| 2036 |
], repo_id
|
| 2037 |
|
| 2038 |
-
|
| 2039 |
-
|
| 2040 |
-
|
| 2041 |
-
|
| 2042 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2043 |
|
| 2044 |
tokens = _get_token_filename(repo_id=repo_id, subfolder=".")
|
| 2045 |
|
|
@@ -2275,6 +2288,11 @@ english_models = {
|
|
| 2275 |
"csukuangfj/wenet-english-model": _get_wenet_model,
|
| 2276 |
}
|
| 2277 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2278 |
more_than_1600_languages_models = {
|
| 2279 |
"csukuangfj/sherpa-onnx-omnilingual-asr-1600-languages-300M-ctc-int8-2025-11-12": _get_sherpa_onnx_omnilingual_asr_models,
|
| 2280 |
"csukuangfj/sherpa-onnx-omnilingual-asr-1600-languages-300M-ctc-2025-11-12": _get_sherpa_onnx_omnilingual_asr_models,
|
|
@@ -2394,6 +2412,7 @@ portuguese_brazlian_models = {
|
|
| 2394 |
|
| 2395 |
|
| 2396 |
all_models = {
|
|
|
|
| 2397 |
**more_than_1600_languages_models,
|
| 2398 |
**twenty_five_languages_models,
|
| 2399 |
**multi_lingual_models,
|
|
@@ -2418,6 +2437,7 @@ all_models = {
|
|
| 2418 |
}
|
| 2419 |
|
| 2420 |
language_to_models = {
|
|
|
|
| 2421 |
"1600+ languages": list(more_than_1600_languages_models.keys()),
|
| 2422 |
"25 European languages": list(twenty_five_languages_models.keys()),
|
| 2423 |
"Multi-lingual (east aisa)": list(multi_lingual_models.keys()),
|
|
|
|
| 178 |
return more_than_1600_languages_models[repo_id](
|
| 179 |
repo_id, decoding_method=decoding_method, num_active_paths=num_active_paths
|
| 180 |
)
|
| 181 |
+
elif repo_id in funsar_nano_31_languages_models:
|
| 182 |
+
return funsar_nano_31_languages_models[repo_id](
|
| 183 |
+
repo_id, decoding_method=decoding_method, num_active_paths=num_active_paths
|
| 184 |
+
)
|
| 185 |
elif repo_id in twenty_five_languages_models:
|
| 186 |
return twenty_five_languages_models[repo_id](
|
| 187 |
repo_id, decoding_method=decoding_method, num_active_paths=num_active_paths
|
|
|
|
| 2035 |
num_active_paths: int,
|
| 2036 |
) -> sherpa_onnx.OfflineRecognizer:
|
| 2037 |
assert repo_id in [
|
| 2038 |
+
"csukuangfj/sherpa-onnx-sense-voice-funasr-nano-int8-2025-12-17",
|
| 2039 |
+
"csukuangfj/sherpa-onnx-sense-voice-funasr-nano-2025-12-17",
|
| 2040 |
"csukuangfj/sherpa-onnx-sense-voice-zh-en-ja-ko-yue-2024-07-17",
|
| 2041 |
"csukuangfj/sherpa-onnx-sense-voice-zh-en-ja-ko-yue-int8-2025-09-09",
|
| 2042 |
], repo_id
|
| 2043 |
|
| 2044 |
+
if repo_id == "csukuangfj/sherpa-onnx-sense-voice-funasr-nano-2025-12-17":
|
| 2045 |
+
nn_model = _get_nn_model_filename(
|
| 2046 |
+
repo_id=repo_id,
|
| 2047 |
+
filename="model.onnx",
|
| 2048 |
+
subfolder=".",
|
| 2049 |
+
)
|
| 2050 |
+
else:
|
| 2051 |
+
nn_model = _get_nn_model_filename(
|
| 2052 |
+
repo_id=repo_id,
|
| 2053 |
+
filename="model.int8.onnx",
|
| 2054 |
+
subfolder=".",
|
| 2055 |
+
)
|
| 2056 |
|
| 2057 |
tokens = _get_token_filename(repo_id=repo_id, subfolder=".")
|
| 2058 |
|
|
|
|
| 2288 |
"csukuangfj/wenet-english-model": _get_wenet_model,
|
| 2289 |
}
|
| 2290 |
|
| 2291 |
+
funsar_nano_31_languages_models = {
|
| 2292 |
+
"csukuangfj/sherpa-onnx-sense-voice-funasr-nano-int8-2025-12-17": _get_sense_voice_pre_trained_model,
|
| 2293 |
+
"csukuangfj/sherpa-onnx-sense-voice-funasr-nano-2025-12-17": _get_sense_voice_pre_trained_model,
|
| 2294 |
+
}
|
| 2295 |
+
|
| 2296 |
more_than_1600_languages_models = {
|
| 2297 |
"csukuangfj/sherpa-onnx-omnilingual-asr-1600-languages-300M-ctc-int8-2025-11-12": _get_sherpa_onnx_omnilingual_asr_models,
|
| 2298 |
"csukuangfj/sherpa-onnx-omnilingual-asr-1600-languages-300M-ctc-2025-11-12": _get_sherpa_onnx_omnilingual_asr_models,
|
|
|
|
| 2412 |
|
| 2413 |
|
| 2414 |
all_models = {
|
| 2415 |
+
**funsar_nano_31_languages_models,
|
| 2416 |
**more_than_1600_languages_models,
|
| 2417 |
**twenty_five_languages_models,
|
| 2418 |
**multi_lingual_models,
|
|
|
|
| 2437 |
}
|
| 2438 |
|
| 2439 |
language_to_models = {
|
| 2440 |
+
"31 languages (FunASR Nano)": list(funsar_nano_31_languages_models.keys()),
|
| 2441 |
"1600+ languages": list(more_than_1600_languages_models.keys()),
|
| 2442 |
"25 European languages": list(twenty_five_languages_models.keys()),
|
| 2443 |
"Multi-lingual (east aisa)": list(multi_lingual_models.keys()),
|
requirements.txt
CHANGED
|
@@ -10,7 +10,7 @@ numpy<2
|
|
| 10 |
|
| 11 |
huggingface_hub
|
| 12 |
|
| 13 |
-
https://huggingface.co/csukuangfj/sherpa-onnx-wheels/resolve/main/cpu/1.12.
|
| 14 |
-
https://huggingface.co/csukuangfj/sherpa-onnx-wheels/resolve/main/cpu/1.12.
|
| 15 |
|
| 16 |
#sherpa-onnx>=1.12.6
|
|
|
|
| 10 |
|
| 11 |
huggingface_hub
|
| 12 |
|
| 13 |
+
https://huggingface.co/csukuangfj/sherpa-onnx-wheels/resolve/main/cpu/1.12.19/sherpa_onnx_core-1.12.19-py3-none-manylinux2014_x86_64.whl
|
| 14 |
+
https://huggingface.co/csukuangfj/sherpa-onnx-wheels/resolve/main/cpu/1.12.19/sherpa_onnx-1.12.19-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl
|
| 15 |
|
| 16 |
#sherpa-onnx>=1.12.6
|