游雁
2024-06-13 c553a8db1712c2a5deeef5bbb68bd1fdf8d61ab7
funasr/models/lcbnet/attention.py
@@ -11,6 +11,7 @@
import torch
from torch import nn
class MultiHeadedAttentionReturnWeight(nn.Module):
    """Multi-Head Attention layer.
@@ -108,5 +109,3 @@
        q, k, v = self.forward_qkv(query, key, value)
        scores = torch.matmul(q, k.transpose(-2, -1)) / math.sqrt(self.d_k)
        return self.forward_attention(v, scores, mask)