From 4ba1011b42e041ee1d71448eefd7ef2e7bd61bb6 Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期五, 31 三月 2023 15:31:26 +0800
Subject: [PATCH] export
---
funasr/bin/modelscope_infer.py | 12 ++++++++++--
1 files changed, 10 insertions(+), 2 deletions(-)
diff --git a/funasr/bin/modelscope_infer.py b/funasr/bin/modelscope_infer.py
index 440c881..bc24340 100755
--- a/funasr/bin/modelscope_infer.py
+++ b/funasr/bin/modelscope_infer.py
@@ -15,6 +15,10 @@
type=str,
default="speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
help="model name in modelscope")
+ parser.add_argument("--model_revision",
+ type=str,
+ default="v1.0.4",
+ help="model revision in modelscope")
parser.add_argument("--local_model_path",
type=str,
default=None,
@@ -62,7 +66,8 @@
if args.local_model_path is None:
inference_pipeline = pipeline(
task=Tasks.auto_speech_recognition,
- model="damo/{}".format(args.model_name))
+ model="damo/{}".format(args.model_name),
+ model_revision=args.model_revision)
else:
inference_pipeline = pipeline(
task=Tasks.auto_speech_recognition,
@@ -77,6 +82,9 @@
wav_id, wav_path = line.strip().split()
logging.info("decoding, utt_id: ['{}']".format(wav_id))
rec_result = inference_pipeline(audio_in=wav_path)
- text = rec_result["text"]
+ if 'text' in rec_result:
+ text = rec_result["text"]
+ else:
+ text = ''
f_out.write(wav_id + " " + text + "\n")
logging.info("best hypo: {} \n".format(text))
--
Gitblit v1.9.1