From f8c740d5a8efe4eeae177c413bdbd51569336ee0 Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期一, 24 七月 2023 16:28:23 +0800
Subject: [PATCH] readme docs

---
 funasr/runtime/onnxruntime/readme.md |  171 --------------------------------------------------------
 1 files changed, 1 insertions(+), 170 deletions(-)

diff --git a/funasr/runtime/onnxruntime/readme.md b/funasr/runtime/onnxruntime/readme.md
index 4ed184f..4865376 100644
--- a/funasr/runtime/onnxruntime/readme.md
+++ b/funasr/runtime/onnxruntime/readme.md
@@ -1,170 +1 @@
-# ONNXRuntime-cpp
-
-## Export the model
-### Install [modelscope and funasr](https://github.com/alibaba-damo-academy/FunASR#installation)
-
-```shell
-# pip3 install torch torchaudio
-pip install -U modelscope funasr
-# For the users in China, you could install with the command:
-# pip install -U modelscope funasr -f https://modelscope.oss-cn-beijing.aliyuncs.com/releases/repo.html -i https://mirror.sjtu.edu.cn/pypi/web/simple
-```
-
-### Export [onnx model](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/export)
-
-```shell
-python -m funasr.export.export_model --model-name damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch --export-dir ./export --type onnx --quantize True
-```
-
-## Building for Linux/Unix
-
-### Download onnxruntime
-```shell
-# download an appropriate onnxruntime from https://github.com/microsoft/onnxruntime/releases/tag/v1.14.0
-# here we get a copy of onnxruntime for linux 64
-wget https://github.com/microsoft/onnxruntime/releases/download/v1.14.0/onnxruntime-linux-x64-1.14.0.tgz
-tar -zxvf onnxruntime-linux-x64-1.14.0.tgz
-```
-
-### Install openblas
-```shell
-sudo apt-get install libopenblas-dev #ubuntu
-# sudo yum -y install openblas-devel #centos
-```
-
-### Build runtime
-```shell
-git clone https://github.com/alibaba-damo-academy/FunASR.git && cd funasr/runtime/onnxruntime
-mkdir build && cd build
-cmake  -DCMAKE_BUILD_TYPE=release .. -DONNXRUNTIME_DIR=/path/to/onnxruntime-linux-x64-1.14.0
-make
-```
-## Run the demo
-
-### funasr-onnx-offline
-```shell
-./funasr-onnx-offline     --model-dir <string> [--quantize <string>]
-                          [--vad-dir <string>] [--vad-quant <string>]
-                          [--punc-dir <string>] [--punc-quant <string>]
-                          --wav-path <string> [--] [--version] [-h]
-Where:
-   --model-dir <string>
-     (required)  the asr model path, which contains model.onnx, config.yaml, am.mvn
-   --quantize <string>
-     false (Default), load the model of model.onnx in model_dir. If set true, load the model of model_quant.onnx in model_dir
-
-   --vad-dir <string>
-     the vad model path, which contains model.onnx, vad.yaml, vad.mvn
-   --vad-quant <string>
-     false (Default), load the model of model.onnx in vad_dir. If set true, load the model of model_quant.onnx in vad_dir
-
-   --punc-dir <string>
-     the punc model path, which contains model.onnx, punc.yaml
-   --punc-quant <string>
-     false (Default), load the model of model.onnx in punc_dir. If set true, load the model of model_quant.onnx in punc_dir
-
-   --wav-path <string>
-     (required)  the input could be: 
-      wav_path, e.g.: asr_example.wav;
-      pcm_path, e.g.: asr_example.pcm; 
-      wav.scp, kaldi style wav list (wav_id \t wav_path)
-  
-   Required: --model-dir <string> --wav-path <string>
-   If use vad, please add: --vad-dir <string>
-   If use punc, please add: --punc-dir <string>
-
-For example:
-./funasr-onnx-offline \
-    --model-dir    ./asrmodel/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch \
-    --quantize  true \
-    --vad-dir   ./asrmodel/speech_fsmn_vad_zh-cn-16k-common-pytorch \
-    --punc-dir  ./asrmodel/punc_ct-transformer_zh-cn-common-vocab272727-pytorch \
-    --wav-path    ./vad_example.wav
-```
-
-### funasr-onnx-offline-vad
-```shell
-./funasr-onnx-offline-vad     --model-dir <string> [--quantize <string>]
-                              --wav-path <string> [--] [--version] [-h]
-Where:
-   --model-dir <string>
-     (required)  the vad model path, which contains model.onnx, vad.yaml, vad.mvn
-   --quantize <string>
-     false (Default), load the model of model.onnx in model_dir. If set true, load the model of model_quant.onnx in model_dir
-   --wav-path <string>
-     (required)  the input could be: 
-      wav_path, e.g.: asr_example.wav;
-      pcm_path, e.g.: asr_example.pcm; 
-      wav.scp, kaldi style wav list (wav_id \t wav_path)
-
-   Required: --model-dir <string> --wav-path <string>
-
-For example:
-./funasr-onnx-offline-vad \
-    --model-dir   ./asrmodel/speech_fsmn_vad_zh-cn-16k-common-pytorch \
-    --wav-path    ./vad_example.wav
-```
-
-### funasr-onnx-offline-punc
-```shell
-./funasr-onnx-offline-punc    --model-dir <string> [--quantize <string>]
-                              --txt-path <string> [--] [--version] [-h]
-Where:
-   --model-dir <string>
-     (required)  the punc model path, which contains model.onnx, punc.yaml
-   --quantize <string>
-     false (Default), load the model of model.onnx in model_dir. If set true, load the model of model_quant.onnx in model_dir
-   --txt-path <string>
-     (required)  txt file path, one sentence per line
-
-   Required: --model-dir <string> --txt-path <string>
-
-For example:
-./funasr-onnx-offline-punc \
-    --model-dir  ./asrmodel/punc_ct-transformer_zh-cn-common-vocab272727-pytorch \
-    --txt-path   ./punc_example.txt
-```
-### funasr-onnx-offline-rtf
-```shell
-./funasr-onnx-offline-rtf     --model-dir <string> [--quantize <string>]
-                              [--vad-dir <string>] [--vad-quant <string>]
-                              [--punc-dir <string>] [--punc-quant <string>]
-                              --wav-path <string> --thread-num <int32_t>
-                              [--] [--version] [-h]
-Where:
-   --thread-num <int32_t>
-     (required)  multi-thread num for rtf
-   --model-dir <string>
-     (required)  the model path, which contains model.onnx, config.yaml, am.mvn
-   --quantize <string>
-     false (Default), load the model of model.onnx in model_dir. If set true, load the model of model_quant.onnx in model_dir
-
-   --vad-dir <string>
-     the vad model path, which contains model.onnx, vad.yaml, vad.mvn
-   --vad-quant <string>
-     false (Default), load the model of model.onnx in vad_dir. If set true, load the model of model_quant.onnx in vad_dir
-
-   --punc-dir <string>
-     the punc model path, which contains model.onnx, punc.yaml
-   --punc-quant <string>
-     false (Default), load the model of model.onnx in punc_dir. If set true, load the model of model_quant.onnx in punc_dir
-     
-   --wav-path <string>
-     (required)  the input could be: 
-      wav_path, e.g.: asr_example.wav;
-      pcm_path, e.g.: asr_example.pcm; 
-      wav.scp, kaldi style wav list (wav_id \t wav_path)
-
-For example:
-./funasr-onnx-offline-rtf \
-    --model-dir    ./asrmodel/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch \
-    --quantize  true \
-    --wav-path     ./aishell1_test.scp  \
-    --thread-num 32
-```
-
-## Acknowledge
-1. This project is maintained by [FunASR community](https://github.com/alibaba-damo-academy/FunASR).
-2. We acknowledge mayong for contributing the onnxruntime of Paraformer and CT_Transformer, [repo-asr](https://github.com/RapidAI/RapidASR/tree/main/cpp_onnx), [repo-punc](https://github.com/RapidAI/RapidPunc).
-3. We acknowledge [ChinaTelecom](https://github.com/zhuzizyf/damo-fsmn-vad-infer-httpserver) for contributing the VAD runtime.
-4. We borrowed a lot of code from [FastASR](https://github.com/chenkui164/FastASR) for audio frontend and text-postprocess.
+Please ref to [websocket service](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/runtime/websocket)
\ No newline at end of file

--
Gitblit v1.9.1