| | |
| | | ## Inference with Triton |
| | | |
| | | ### Steps: |
| | | 1. Refer here to [get model.onnx](https://github.com/alibaba-damo-academy/FunASR/tree/main/funasr/runtime/python/onnxruntime#steps) |
| | | 1. Refer here to [get model.onnx](https://github.com/alibaba-damo-academy/FunASR/blob/main/funasr/export/README.md) |
| | | |
| | | 2. Follow below instructions to using triton |
| | | ```sh |
| | |
| | | | 60 (onnx fp32) | 116.0 | 0.0032| |
| | | |
| | | ## Acknowledge |
| | | This part originates from NVIDIA CISI project. We also have TTS and NLP solutions deployed on triton inference server. If you are interested, please contact us. |
| | | This part originates from NVIDIA CISI project. We also have TTS and NLP solutions deployed on triton inference server. If you are interested, please contact us. |