Shi Xian
2025-02-18 57e023e5cfa522007e24d87c8e3d82bf7c4a19cd
funasr/models/ct_transformer_streaming/attention.py
@@ -18,5 +18,3 @@
        scores = torch.matmul(q_h, k_h.transpose(-2, -1))
        att_outs = self.forward_attention(v_h, scores, mask[1], mask_att_chunk_encoder)
        return att_outs + fsmn_memory