From 2ac38adbe5f4e1374a079e032ed4b504351a207c Mon Sep 17 00:00:00 2001
From: zhifu gao <zhifu.gzf@alibaba-inc.com>
Date: 星期二, 23 四月 2024 18:08:57 +0800
Subject: [PATCH] Dev gzf exp (#1647)
---
funasr/models/sense_voice/decoder.py | 24 +-----------------------
1 files changed, 1 insertions(+), 23 deletions(-)
diff --git a/funasr/models/sense_voice/decoder.py b/funasr/models/sense_voice/decoder.py
index 9087ea1..9fdb3bd 100644
--- a/funasr/models/sense_voice/decoder.py
+++ b/funasr/models/sense_voice/decoder.py
@@ -245,29 +245,7 @@
self.register_buffer("mask", mask, persistent=False)
self.use_padmask = kwargs.get("use_padmask", True)
- # def forward(self, x: Tensor, xa: Tensor, kv_cache: Optional[dict] = None):
- # """
- # x : torch.LongTensor, shape = (batch_size, <= n_ctx)
- # the text tokens
- # xa : torch.Tensor, shape = (batch_size, n_audio_ctx, n_audio_state)
- # the encoded audio features to be attended on
- # """
- # offset = next(iter(kv_cache.values())).shape[1] if kv_cache else 0
- # x = (
- # self.token_embedding(x)
- # + self.positional_embedding[offset: offset + x.shape[-1]]
- # )
- # x = x.to(xa.dtype)
- #
- # for block in self.blocks:
- # x = block(x, xa, mask=self.mask, kv_cache=kv_cache)
- #
- # x = self.ln(x)
- # logits = (
- # x @ torch.transpose(self.token_embedding.weight.to(x.dtype), 0, 1)
- # ).float()
- #
- # return logits
+
def forward(
--
Gitblit v1.9.1