egs_modelscope/asr/uniasr/speech_UniASR_asr_2pass-zh-cn-8k-common-vocab3445-pytorch-offline/infer.py
@@ -23,8 +23,7 @@ batch_size=1 ) audio_in = os.path.join(split_dir, "wav.{}.scp".format(idx)) inference_pipline(audio_in=audio_in, param_dict={"decoding_model": "offline"}) inference_pipline(audio_in=audio_in) def modelscope_infer(params): # prepare for multi-GPU decoding