From ed3e085f78c6d618eaf6bc78eb11ea9d77a1e298 Mon Sep 17 00:00:00 2001 From: 游雁 <zhifu.gzf@alibaba-inc.com> Date: 星期六, 22 四月 2023 09:50:46 +0800 Subject: [PATCH] infer.sh --- egs_modelscope/asr/TEMPLATE/README.md | 4 ++-- 1 files changed, 2 insertions(+), 2 deletions(-) diff --git a/egs_modelscope/asr/TEMPLATE/README.md b/egs_modelscope/asr/TEMPLATE/README.md index 7e81f87..19acefe 100644 --- a/egs_modelscope/asr/TEMPLATE/README.md +++ b/egs_modelscope/asr/TEMPLATE/README.md @@ -53,7 +53,7 @@ rec_result = inference_pipeline(audio_in='https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav') print(rec_result) ``` -The decoding mode of `fast` and `normal` +The decoding mode of `fast` and `normal` is fake streaming, which could be used for evaluating of recognition accuracy. Full code of demo, please ref to [demo](https://github.com/alibaba-damo-academy/FunASR/discussions/151) #### [RNN-T-online model]() Undo @@ -186,7 +186,7 @@ ``` ## Inference with your finetuned model -- Setting parameters in [infer.sh](https://github.com/alibaba-damo-academy/FunASR/blob/main/egs_modelscope/asr/TEMPLATE/infer.sh) is the same with [docs](https://github.com/alibaba-damo-academy/FunASR/tree/main/egs_modelscope/asr/TEMPLATE#inference-with-multi-thread-cpus-or-multi-gpus) +- Setting parameters in [infer.sh](https://github.com/alibaba-damo-academy/FunASR/blob/main/egs_modelscope/asr/TEMPLATE/infer.sh) is the same with [docs](https://github.com/alibaba-damo-academy/FunASR/tree/main/egs_modelscope/asr/TEMPLATE#inference-with-multi-thread-cpus-or-multi-gpus), `model` is the model name from modelscope, which you finetuned. - Decode with multi GPUs: ```shell -- Gitblit v1.9.1