aky15
2023-05-17 9d01231fa69672fc7c9b4bf81ef466bb0189788c
funasr/models/encoder/conformer_encoder.py
@@ -1078,7 +1078,7 @@
                limit_size,
            )
        mask = make_source_mask(x_len)
        mask = make_source_mask(x_len).to(x.device)
        if self.unified_model_training:
            chunk_size = self.default_chunk_size + torch.randint(-self.jitter_range, self.jitter_range+1, (1,)).item()