游雁
2023-04-07 9f6445d39b14fa17f2c32a383c1054a8e073a9c9
funasr/export/models/modules/multihead_att.py
@@ -75,8 +75,10 @@
    return x, cache
import torch.fx
torch.fx.wrap('preprocess_for_attn')
torch_version = float(".".join(torch.__version__.split(".")[:2]))
if torch_version >= 1.8:
    import torch.fx
    torch.fx.wrap('preprocess_for_attn')
class MultiHeadedAttentionSANMDecoder(nn.Module):