From a513d1ac95ae6f7d7cbc6dfefc9bbb3f78718163 Mon Sep 17 00:00:00 2001
From: Lizerui9926 <110582652+Lizerui9926@users.noreply.github.com>
Date: 星期二, 04 四月 2023 14:35:36 +0800
Subject: [PATCH] Merge pull request #325 from alibaba-damo-academy/dev_lzr
---
funasr/models/encoder/opennmt_encoders/self_attention_encoder.py | 2 +-
1 files changed, 1 insertions(+), 1 deletions(-)
diff --git a/funasr/models/encoder/opennmt_encoders/self_attention_encoder.py b/funasr/models/encoder/opennmt_encoders/self_attention_encoder.py
index 443b37a..cf77bce 100644
--- a/funasr/models/encoder/opennmt_encoders/self_attention_encoder.py
+++ b/funasr/models/encoder/opennmt_encoders/self_attention_encoder.py
@@ -272,7 +272,7 @@
position embedded tensor and mask
"""
masks = (~make_pad_mask(ilens)[:, None, :]).to(xs_pad.device)
- xs_pad *= self.output_size()**0.5
+ xs_pad = xs_pad * self.output_size()**0.5
if self.embed is None:
xs_pad = xs_pad
elif (
--
Gitblit v1.9.1