| | |
| | | |
| | | import numpy as np |
| | | import torch |
| | | from typeguard import check_argument_types |
| | | from funasr.build_utils.build_streaming_iterator import build_streaming_iterator |
| | | from funasr.fileio.datadir_writer import DatadirWriter |
| | | from funasr.torch_utils.set_all_random_seed import set_all_random_seed |
| | |
| | | num_workers: int = 1, |
| | | **kwargs, |
| | | ): |
| | | assert check_argument_types() |
| | | if batch_size > 1: |
| | | raise NotImplementedError("batch decoding is not implemented") |
| | | |
| | |
| | | vad_results.append(item) |
| | | if writer is not None: |
| | | ibest_writer["text"][keys[i]] = "{}".format(results[i]) |
| | | |
| | | torch.cuda.empty_cache() |
| | | return vad_results |
| | | |
| | | return _forward |
| | |
| | | num_workers: int = 1, |
| | | **kwargs, |
| | | ): |
| | | assert check_argument_types() |
| | | |
| | | logging.basicConfig( |
| | | level=log_level, |