From 4853dffdc4e70552f5b4348ef29089ed04927d95 Mon Sep 17 00:00:00 2001 From: zhifu gao <zhifu.gzf@alibaba-inc.com> Date: 星期五, 17 二月 2023 11:43:40 +0800 Subject: [PATCH] Merge pull request #131 from alibaba-damo-academy/dev_lyb --- README.md | 2 +- 1 files changed, 1 insertions(+), 1 deletions(-) diff --git a/README.md b/README.md index 991941b..eaeb9ba 100644 --- a/README.md +++ b/README.md @@ -17,7 +17,7 @@ ## What's new: -### 2023.2.16, funasr-0.2.0, modelscope-1.3.0 +### 2023.2.17, funasr-0.2.0, modelscope-1.3.0 - We support a new feature, export paraformer models into [onnx and torchscripts](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/export) from modelscope. The local finetuned models are also supported. - We support a new feature, [onnxruntime](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/runtime/python/onnxruntime/paraformer/rapid_paraformer), you could deploy the runtime without modelscope or funasr, for the [paraformer-large](https://www.modelscope.cn/models/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/summary) model, the rtf of onnxruntime is 3x speedup(0.110->0.038) on cpu, [details](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/runtime/python/onnxruntime/paraformer/rapid_paraformer#speed). - We support a new feature, [grpc](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/runtime/python/grpc), you could build the ASR service with grpc, by deploying the modelscope pipeline or onnxruntime. -- Gitblit v1.9.1