aky15
2023-05-23 71f1059af9bd71ce87483952bbf3964fc3d3a5f9
funasr/models/encoder/conformer_encoder.py
@@ -1078,7 +1078,7 @@
                limit_size,
            )
        mask = make_source_mask(x_len)
        mask = make_source_mask(x_len).to(x.device)
        if self.unified_model_training:
            chunk_size = self.default_chunk_size + torch.randint(-self.jitter_range, self.jitter_range+1, (1,)).item()