| | |
| | | self.streaming = streaming |
| | | self.simu_streaming = simu_streaming |
| | | self.chunk_size = max(chunk_size, 0) |
| | | self.left_context = max(left_context, 0) |
| | | self.left_context = left_context |
| | | self.right_context = max(right_context, 0) |
| | | |
| | | if not streaming or chunk_size == 0: |
| | |
| | | sym_blank: str = "<blank>", |
| | | report_cer: bool = True, |
| | | report_wer: bool = True, |
| | | sym_sos: str = "<sos/eos>", |
| | | sym_eos: str = "<sos/eos>", |
| | | sym_sos: str = "<s>", |
| | | sym_eos: str = "</s>", |
| | | extract_feats_in_collect_stats: bool = True, |
| | | lsm_weight: float = 0.0, |
| | | length_normalized_loss: bool = False, |
| | |
| | | mask = torch.zeros(size, size, device=device, dtype=torch.bool) |
| | | |
| | | for i in range(size): |
| | | if left_chunk_size <= 0: |
| | | if left_chunk_size < 0: |
| | | start = 0 |
| | | else: |
| | | start = max((i // chunk_size - left_chunk_size) * chunk_size, 0) |