funasr/models/encoder/opennmt_encoders/self_attention_encoder.py
@@ -117,7 +117,7 @@ class SelfAttentionEncoder(AbsEncoder): """ author: Speech Lab, Alibaba Group, China Author: Speech Lab of DAMO Academy, Alibaba Group Self attention encoder in OpenNMT framework """ @@ -272,7 +272,7 @@ position embedded tensor and mask """ masks = (~make_pad_mask(ilens)[:, None, :]).to(xs_pad.device) xs_pad *= self.output_size()**0.5 xs_pad = xs_pad * self.output_size()**0.5 if self.embed is None: xs_pad = xs_pad elif (