From 91027ddab49e5791fc42569b4db9dafca55735e6 Mon Sep 17 00:00:00 2001
From: 凌匀 <ailsa.zly@alibaba-inc.com>
Date: 星期四, 16 二月 2023 22:11:18 +0800
Subject: [PATCH] fix vad results bug
---
funasr/bin/vad_inference.py | 132 +++++++++++--------------------------------
1 files changed, 34 insertions(+), 98 deletions(-)
diff --git a/funasr/bin/vad_inference.py b/funasr/bin/vad_inference.py
index 679cc0b..b0f8a77 100644
--- a/funasr/bin/vad_inference.py
+++ b/funasr/bin/vad_inference.py
@@ -1,6 +1,7 @@
import argparse
import logging
import sys
+import json
from pathlib import Path
from typing import Any
from typing import List
@@ -80,6 +81,7 @@
self.device = device
self.dtype = dtype
self.frontend = frontend
+ self.batch_size = batch_size
@torch.no_grad()
def __call__(
@@ -105,101 +107,31 @@
feats_len = feats_len.int()
else:
raise Exception("Need to extract feats first, please configure frontend configuration")
- batch = {"feats": feats, "feats_lengths": feats_len, "waveform": speech}
+ # batch = {"feats": feats, "waveform": speech, "is_final_send": True}
+ # segments = self.vad_model(**batch)
- # a. To device
- batch = to_device(batch, device=self.device)
-
- # b. Forward Encoder
- segments = self.vad_model(**batch)
-
+ # b. Forward Encoder sreaming
+ t_offset = 0
+ step = min(feats_len, 6000)
+ segments = [[]] * self.batch_size
+ for t_offset in range(0, feats_len, min(step, feats_len - t_offset)):
+ if t_offset + step >= feats_len - 1:
+ step = feats_len - t_offset
+ is_final_send = True
+ else:
+ is_final_send = False
+ batch = {
+ "feats": feats[:, t_offset:t_offset + step, :],
+ "waveform": speech[:, t_offset * 160:min(speech.shape[-1], (t_offset + step - 1) * 160 + 400)],
+ "is_final_send": is_final_send
+ }
+ # a. To device
+ batch = to_device(batch, device=self.device)
+ segments_part = self.vad_model(**batch)
+ if segments_part:
+ for batch_num in range(0, self.batch_size):
+ segments[batch_num] += segments_part[batch_num]
return segments
-
-
-#def inference(
-# batch_size: int,
-# ngpu: int,
-# log_level: Union[int, str],
-# data_path_and_name_and_type,
-# vad_infer_config: Optional[str],
-# vad_model_file: Optional[str],
-# vad_cmvn_file: Optional[str] = None,
-# raw_inputs: Union[np.ndarray, torch.Tensor] = None,
-# key_file: Optional[str] = None,
-# allow_variable_data_keys: bool = False,
-# output_dir: Optional[str] = None,
-# dtype: str = "float32",
-# seed: int = 0,
-# num_workers: int = 1,
-# fs: Union[dict, int] = 16000,
-# **kwargs,
-#):
-# assert check_argument_types()
-# if batch_size > 1:
-# raise NotImplementedError("batch decoding is not implemented")
-# if ngpu > 1:
-# raise NotImplementedError("only single GPU decoding is supported")
-#
-# logging.basicConfig(
-# level=log_level,
-# format="%(asctime)s (%(module)s:%(lineno)d) %(levelname)s: %(message)s",
-# )
-#
-# if ngpu >= 1 and torch.cuda.is_available():
-# device = "cuda"
-# else:
-# device = "cpu"
-#
-# # 1. Set random-seed
-# set_all_random_seed(seed)
-#
-# # 2. Build speech2vadsegment
-# speech2vadsegment_kwargs = dict(
-# vad_infer_config=vad_infer_config,
-# vad_model_file=vad_model_file,
-# vad_cmvn_file=vad_cmvn_file,
-# device=device,
-# dtype=dtype,
-# )
-# logging.info("speech2vadsegment_kwargs: {}".format(speech2vadsegment_kwargs))
-# speech2vadsegment = Speech2VadSegment(**speech2vadsegment_kwargs)
-# # 3. Build data-iterator
-# loader = VADTask.build_streaming_iterator(
-# data_path_and_name_and_type,
-# dtype=dtype,
-# batch_size=batch_size,
-# key_file=key_file,
-# num_workers=num_workers,
-# preprocess_fn=VADTask.build_preprocess_fn(speech2vadsegment.vad_infer_args, False),
-# collate_fn=VADTask.build_collate_fn(speech2vadsegment.vad_infer_args, False),
-# allow_variable_data_keys=allow_variable_data_keys,
-# inference=True,
-# )
-#
-# finish_count = 0
-# file_count = 1
-# # 7 .Start for-loop
-# # FIXME(kamo): The output format should be discussed about
-# if output_dir is not None:
-# writer = DatadirWriter(output_dir)
-# else:
-# writer = None
-#
-# vad_results = []
-# for keys, batch in loader:
-# assert isinstance(batch, dict), type(batch)
-# assert all(isinstance(s, str) for s in keys), keys
-# _bs = len(next(iter(batch.values())))
-# assert len(keys) == _bs, f"{len(keys)} != {_bs}"
-# # batch = {k: v[0] for k, v in batch.items() if not k.endswith("_lengths")}
-#
-# # do vad segment
-# results = speech2vadsegment(**batch)
-# for i, _ in enumerate(keys):
-# item = {'key': keys[i], 'value': results[i]}
-# vad_results.append(item)
-#
-# return vad_results
def inference(
@@ -236,11 +168,12 @@
)
return inference_pipeline(data_path_and_name_and_type, raw_inputs)
+
def inference_modelscope(
batch_size: int,
ngpu: int,
log_level: Union[int, str],
- #data_path_and_name_and_type,
+ # data_path_and_name_and_type,
vad_infer_config: Optional[str],
vad_model_file: Optional[str],
vad_cmvn_file: Optional[str] = None,
@@ -284,9 +217,11 @@
speech2vadsegment = Speech2VadSegment(**speech2vadsegment_kwargs)
def _forward(
- data_path_and_name_and_type,
- raw_inputs: Union[np.ndarray, torch.Tensor] = None,
- output_dir_v2: Optional[str] = None,
+ data_path_and_name_and_type,
+ raw_inputs: Union[np.ndarray, torch.Tensor] = None,
+ output_dir_v2: Optional[str] = None,
+ fs: dict = None,
+ param_dict: dict = None,
):
# 3. Build data-iterator
loader = VADTask.build_streaming_iterator(
@@ -319,14 +254,15 @@
assert all(isinstance(s, str) for s in keys), keys
_bs = len(next(iter(batch.values())))
assert len(keys) == _bs, f"{len(keys)} != {_bs}"
- # batch = {k: v[0] for k, v in batch.items() if not k.endswith("_lengths")}
# do vad segment
results = speech2vadsegment(**batch)
for i, _ in enumerate(keys):
+ results[i] = json.dumps(results[i])
item = {'key': keys[i], 'value': results[i]}
vad_results.append(item)
if writer is not None:
+ results[i] = json.loads(results[i])
ibest_writer["text"][keys[i]] = "{}".format(results[i])
return vad_results
--
Gitblit v1.9.1