From 828b5f81afd5173cda6eec980017b28d0b298069 Mon Sep 17 00:00:00 2001 From: 游雁 <zhifu.gzf@alibaba-inc.com> Date: 星期二, 18 四月 2023 13:04:51 +0800 Subject: [PATCH] docs --- funasr/runtime/python/onnxruntime/README.md | 63 +++++++++++++++---------------- 1 files changed, 30 insertions(+), 33 deletions(-) diff --git a/funasr/runtime/python/onnxruntime/README.md b/funasr/runtime/python/onnxruntime/README.md index 87510fa..e85e08a 100644 --- a/funasr/runtime/python/onnxruntime/README.md +++ b/funasr/runtime/python/onnxruntime/README.md @@ -1,30 +1,28 @@ -## Using funasr with ONNXRuntime +# ONNXRuntime-python + +## Export the model +### Install [modelscope and funasr](https://github.com/alibaba-damo-academy/FunASR#installation) + +```shell +pip3 install torch torchaudio +pip install -U modelscope +pip install -U funasr +``` + +### Export [onnx model](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/export) + +```shell +python -m funasr.export.export_model --model-name damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch --export-dir ./export --type onnx --quantize True +``` -### Steps: -1. Export the model. - - Command: (`Tips`: torch >= 1.11.0 is required.) - - More details ref to ([export docs](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/export)) - - - `e.g.`, Export model from modelscope - ```shell - python -m funasr.export.export_model --model-name damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch --export-dir ./export --type onnx --quantize False - ``` - - `e.g.`, Export model from local path, the model'name must be `model.pb`. - ```shell - python -m funasr.export.export_model --model-name ./damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch --export-dir ./export --type onnx --quantize False - ``` - - -2. Install the `funasr_onnx` +## Install the `funasr_onnx` install from pip ```shell pip install -U funasr_onnx # For the users in China, you could install with the command: # pip install -U funasr_onnx -i https://mirror.sjtu.edu.cn/pypi/web/simple - ``` or install from source code @@ -35,25 +33,24 @@ pip install -e ./ # For the users in China, you could install with the command: # pip install -e ./ -i https://mirror.sjtu.edu.cn/pypi/web/simple - ``` -3. Run the demo. - - Model_dir: the model path, which contains `model.onnx`, `config.yaml`, `am.mvn`. - - Input: wav formt file, support formats: `str, np.ndarray, List[str]` - - Output: `List[str]`: recognition result. - - Example: - ```python - from funasr_onnx import Paraformer +## Run the demo +- Model_dir: the model path, which contains `model.onnx`, `config.yaml`, `am.mvn`. +- Input: wav formt file, support formats: `str, np.ndarray, List[str]` +- Output: `List[str]`: recognition result. +- Example: + ```python + from funasr_onnx import Paraformer - model_dir = "/nfs/zhifu.gzf/export/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch" - model = Paraformer(model_dir, batch_size=1) + model_dir = "/nfs/zhifu.gzf/export/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch" + model = Paraformer(model_dir, batch_size=1) - wav_path = ['/nfs/zhifu.gzf/export/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/example/asr_example.wav'] + wav_path = ['/nfs/zhifu.gzf/export/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/example/asr_example.wav'] - result = model(wav_path) - print(result) - ``` + result = model(wav_path) + print(result) + ``` ## Performance benchmark -- Gitblit v1.9.1