1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
| #!/usr/bin/env python3
| # -*- encoding: utf-8 -*-
| # Copyright FunASR (https://github.com/alibaba-damo-academy/FunASR). All Rights Reserved.
| # MIT License (https://opensource.org/licenses/MIT)
|
| # To install requirements: pip3 install -U openai-whisper
|
| from funasr import AutoModel
|
| model = AutoModel(
| model="Whisper-large-v3-turbo",
| vad_model="iic/speech_fsmn_vad_zh-cn-16k-common-pytorch",
| vad_kwargs={"max_single_segment_time": 30000},
| )
|
| DecodingOptions = {
| "task": "transcribe",
| "language": None,
| "beam_size": None,
| "fp16": True,
| "without_timestamps": False,
| "prompt": None,
| }
| res = model.generate(
| DecodingOptions=DecodingOptions,
| batch_size_s=0,
| input="https://isv-data.oss-cn-hangzhou.aliyuncs.com/ics/MaaS/ASR/test_audio/asr_example_zh.wav",
| )
|
| print(res)
|
|