From 9b4e9cc8a0311e5243d69b73ed073e7ea441982e Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期三, 27 三月 2024 16:05:29 +0800
Subject: [PATCH] train update
---
funasr/frontends/wav_frontend.py | 11 ++++++-----
1 files changed, 6 insertions(+), 5 deletions(-)
diff --git a/funasr/frontends/wav_frontend.py b/funasr/frontends/wav_frontend.py
index fe22335..afa7421 100644
--- a/funasr/frontends/wav_frontend.py
+++ b/funasr/frontends/wav_frontend.py
@@ -75,6 +75,7 @@
LFR_outputs = torch.vstack(LFR_inputs)
return LFR_outputs.type(torch.float32)
+@tables.register("frontend_classes", "wav_frontend")
@tables.register("frontend_classes", "WavFrontend")
class WavFrontend(nn.Module):
"""Conventional frontend structure for ASR.
@@ -399,11 +400,11 @@
return feats_pad, feats_lens, lfr_splice_frame_idxs
def forward(
- self, input: torch.Tensor, input_lengths: torch.Tensor, cache: dict = {}, **kwargs
+ self, input: torch.Tensor, input_lengths: torch.Tensor, **kwargs
):
is_final = kwargs.get("is_final", False)
- reset = kwargs.get("reset", False)
- if len(cache) == 0 or reset:
+ cache = kwargs.get("cache", {})
+ if len(cache) == 0:
self.init_cache(cache)
batch_size = input.shape[0]
@@ -448,8 +449,8 @@
feats = torch.stack(cache["lfr_splice_cache"])
feats_lengths = torch.zeros(batch_size, dtype=torch.int) + feats.shape[1]
feats, feats_lengths, _ = self.forward_lfr_cmvn(feats, feats_lengths, is_final, cache=cache)
- if is_final:
- self.init_cache(cache)
+ # if is_final:
+ # self.init_cache(cache)
return feats, feats_lengths
--
Gitblit v1.9.1