From 9038340be707aa7cbf62fcbf33ab615bb266abdb Mon Sep 17 00:00:00 2001
From: 王梦迪 <73778524+di-osc@users.noreply.github.com>
Date: 星期一, 26 五月 2025 14:11:02 +0800
Subject: [PATCH] 修复Fsmn_vad_online多线程调用报错 (#2528)

---
 runtime/python/onnxruntime/funasr_onnx/vad_bin.py |   27 ++++++++++++++-------------
 1 files changed, 14 insertions(+), 13 deletions(-)

diff --git a/runtime/python/onnxruntime/funasr_onnx/vad_bin.py b/runtime/python/onnxruntime/funasr_onnx/vad_bin.py
index af4663a..f784f26 100644
--- a/runtime/python/onnxruntime/funasr_onnx/vad_bin.py
+++ b/runtime/python/onnxruntime/funasr_onnx/vad_bin.py
@@ -4,7 +4,7 @@
 
 import os.path
 from pathlib import Path
-from typing import List, Union, Tuple
+from typing import List, Union, Tuple, Dict
 
 import copy
 import librosa
@@ -247,19 +247,17 @@
             model_dir = model.export(type="onnx", quantize=quantize, **kwargs)
 
         config_file = os.path.join(model_dir, "config.yaml")
-        cmvn_file = os.path.join(model_dir, "am.mvn")
-        config = read_yaml(config_file)
+        self.cmvn_file = os.path.join(model_dir, "am.mvn")
+        self.config = read_yaml(config_file)
 
-        self.frontend = WavFrontendOnline(cmvn_file=cmvn_file, **config["frontend_conf"])
         self.ort_infer = OrtInferSession(
             model_file, device_id, intra_op_num_threads=intra_op_num_threads
         )
         self.batch_size = batch_size
-        self.vad_scorer = E2EVadModel(config["model_conf"])
         self.max_end_sil = (
-            max_end_sil if max_end_sil is not None else config["model_conf"]["max_end_silence_time"]
+            max_end_sil if max_end_sil is not None else self.config["model_conf"]["max_end_silence_time"]
         )
-        self.encoder_conf = config["encoder_conf"]
+        self.encoder_conf = self.config["encoder_conf"]
 
     def prepare_cache(self, in_cache: list = []):
         if len(in_cache) > 0:
@@ -275,20 +273,22 @@
     def __call__(self, audio_in: np.ndarray, **kwargs) -> List:
         waveforms = np.expand_dims(audio_in, axis=0)
 
-        param_dict = kwargs.get("param_dict", dict())
+        param_dict: Dict = kwargs.get("param_dict", dict())
         is_final = param_dict.get("is_final", False)
-        feats, feats_len = self.extract_feat(waveforms, is_final)
+        frontend: WavFrontendOnline = param_dict.get("frontend", WavFrontendOnline(cmvn_file=self.cmvn_file, **self.config["frontend_conf"]))
+        feats, feats_len = self.extract_feat(frontend=frontend, waveforms=waveforms, is_final=is_final)
         segments = []
         if feats.size != 0:
             in_cache = param_dict.get("in_cache", list())
             in_cache = self.prepare_cache(in_cache)
+            vad_scorer = param_dict.get("vad_scorer", E2EVadModel(self.config["model_conf"]))
             try:
                 inputs = [feats]
                 inputs.extend(in_cache)
                 scores, out_caches = self.infer(inputs)
                 param_dict["in_cache"] = out_caches
-                waveforms = self.frontend.get_waveforms()
-                segments = self.vad_scorer(
+                waveforms = frontend.get_waveforms()
+                segments = vad_scorer(
                     scores, waveforms, is_final=is_final, max_end_sil=self.max_end_sil, online=True
                 )
 
@@ -296,6 +296,7 @@
                 # logging.warning(traceback.format_exc())
                 logging.warning("input wav is silence or noise")
                 segments = []
+        param_dict.update({"frontend": frontend, "vad_scorer": vad_scorer})
         return segments
 
     def load_data(self, wav_content: Union[str, np.ndarray, List[str]], fs: int = None) -> List:
@@ -315,13 +316,13 @@
         raise TypeError(f"The type of {wav_content} is not in [str, np.ndarray, list]")
 
     def extract_feat(
-        self, waveforms: np.ndarray, is_final: bool = False
+        self, frontend: WavFrontendOnline, waveforms: np.ndarray, is_final: bool = False
     ) -> Tuple[np.ndarray, np.ndarray]:
         waveforms_lens = np.zeros(waveforms.shape[0]).astype(np.int32)
         for idx, waveform in enumerate(waveforms):
             waveforms_lens[idx] = waveform.shape[-1]
 
-        feats, feats_len = self.frontend.extract_fbank(waveforms, waveforms_lens, is_final)
+        feats, feats_len = frontend.extract_fbank(waveforms, waveforms_lens, is_final)
         # feats.append(feat)
         # feats_len.append(feat_len)
 

--
Gitblit v1.9.1