From fdafd3f6bc2f04d16e7cab5afcdb1257e87a8a78 Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期二, 17 十二月 2024 11:15:53 +0800
Subject: [PATCH] emotion2vec

---
 examples/industrial_data_pretraining/qwen_audio/demo_chat.py |    5 ++---
 1 files changed, 2 insertions(+), 3 deletions(-)

diff --git a/examples/industrial_data_pretraining/qwen_audio/demo_chat.py b/examples/industrial_data_pretraining/qwen_audio/demo_chat.py
index a2fb05b..26b5b89 100644
--- a/examples/industrial_data_pretraining/qwen_audio/demo_chat.py
+++ b/examples/industrial_data_pretraining/qwen_audio/demo_chat.py
@@ -7,12 +7,12 @@
 
 from funasr import AutoModel
 
-model = AutoModel(model="Qwen-Audio-Chat")
+model = AutoModel(model="Qwen/Qwen-Audio-Chat")
 
 audio_in = "https://github.com/QwenLM/Qwen-Audio/raw/main/assets/audio/1272-128104-0000.flac"
 
 # 1st dialogue turn
-prompt = 'what does the person say?'
+prompt = "what does the person say?"
 cache = {"history": None}
 res = model.generate(input=audio_in, prompt=prompt, cache=cache)
 print(res)
@@ -22,4 +22,3 @@
 prompt = 'Find the start time and end time of the word "middle classes"'
 res = model.generate(input=None, prompt=prompt, cache=cache)
 print(res)
-

--
Gitblit v1.9.1