funasr/bin/tp_inference.py
@@ -54,7 +54,7 @@ assert check_argument_types() # 1. Build ASR model tp_model, tp_train_args = ASRTask.build_model_from_file( timestamp_infer_config, timestamp_model_file, device timestamp_infer_config, timestamp_model_file, device=device ) if 'cuda' in device: tp_model = tp_model.cuda() # force model to cuda @@ -179,6 +179,9 @@ **kwargs, ): assert check_argument_types() ncpu = kwargs.get("ncpu", 1) torch.set_num_threads(ncpu) if batch_size > 1: raise NotImplementedError("batch decoding is not implemented") if ngpu > 1: