Merge pull request #441 from alibaba-damo-academy/dev_lhn
update
| | |
| | | results = [] |
| | | cache_en = cache["encoder"] |
| | | if speech.shape[1] < 16 * 60 and cache_en["is_final"]: |
| | | if cache_en["start_idx"] == 0: |
| | | return [] |
| | | cache_en["tail_chunk"] = True |
| | | feats = cache_en["feats"] |
| | | feats_len = torch.tensor([feats.shape[1]]) |
| | | self.asr_model.frontend = None |
| | | results = self.infer(feats, feats_len, cache) |
| | | return results |
| | | else: |
| | |
| | | else: |
| | | xs_pad = self.embed(xs_pad, cache) |
| | | if cache["tail_chunk"]: |
| | | xs_pad = cache["feats"] |
| | | xs_pad = to_device(cache["feats"], device=xs_pad.device) |
| | | else: |
| | | xs_pad = self._add_overlap_chunk(xs_pad, cache) |
| | | encoder_outs = self.encoders0(xs_pad, None, None, None, None) |