| | |
| | | type=str, |
| | | default="speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch", |
| | | help="model name in modelscope") |
| | | parser.add_argument("--model_revision", |
| | | type=str, |
| | | default="v1.0.4", |
| | | help="model revision in modelscope") |
| | | parser.add_argument("--local_model_path", |
| | | type=str, |
| | | default=None, |
| | |
| | | if args.local_model_path is None: |
| | | inference_pipeline = pipeline( |
| | | task=Tasks.auto_speech_recognition, |
| | | model="damo/{}".format(args.model_name)) |
| | | model="damo/{}".format(args.model_name), |
| | | model_revision=args.model_revision) |
| | | else: |
| | | inference_pipeline = pipeline( |
| | | task=Tasks.auto_speech_recognition, |
| | |
| | | wav_id, wav_path = line.strip().split() |
| | | logging.info("decoding, utt_id: ['{}']".format(wav_id)) |
| | | rec_result = inference_pipeline(audio_in=wav_path) |
| | | text = rec_result["text"] |
| | | if 'text' in rec_result: |
| | | text = rec_result["text"] |
| | | else: |
| | | text = '' |
| | | f_out.write(wav_id + " " + text + "\n") |
| | | logging.info("best hypo: {} \n".format(text)) |