From 9b4e9cc8a0311e5243d69b73ed073e7ea441982e Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期三, 27 三月 2024 16:05:29 +0800
Subject: [PATCH] train update

---
 funasr/frontends/wav_frontend.py |   11 ++++++-----
 1 files changed, 6 insertions(+), 5 deletions(-)

diff --git a/funasr/frontends/wav_frontend.py b/funasr/frontends/wav_frontend.py
index fe22335..afa7421 100644
--- a/funasr/frontends/wav_frontend.py
+++ b/funasr/frontends/wav_frontend.py
@@ -75,6 +75,7 @@
     LFR_outputs = torch.vstack(LFR_inputs)
     return LFR_outputs.type(torch.float32)
 
+@tables.register("frontend_classes", "wav_frontend")
 @tables.register("frontend_classes", "WavFrontend")
 class WavFrontend(nn.Module):
     """Conventional frontend structure for ASR.
@@ -399,11 +400,11 @@
         return feats_pad, feats_lens, lfr_splice_frame_idxs
 
     def forward(
-        self, input: torch.Tensor, input_lengths: torch.Tensor, cache: dict = {}, **kwargs
+        self, input: torch.Tensor, input_lengths: torch.Tensor, **kwargs
     ):
         is_final = kwargs.get("is_final", False)
-        reset = kwargs.get("reset", False)
-        if len(cache) == 0 or reset:
+        cache = kwargs.get("cache", {})
+        if len(cache) == 0:
             self.init_cache(cache)
         
         batch_size = input.shape[0]
@@ -448,8 +449,8 @@
                 feats = torch.stack(cache["lfr_splice_cache"])
                 feats_lengths = torch.zeros(batch_size, dtype=torch.int) + feats.shape[1]
                 feats, feats_lengths, _ = self.forward_lfr_cmvn(feats, feats_lengths, is_final, cache=cache)
-        if is_final:
-            self.init_cache(cache)
+        # if is_final:
+        #     self.init_cache(cache)
         return feats, feats_lengths
 
 

--
Gitblit v1.9.1