| | |
| | | kwargs = download_model(**kwargs) |
| | | |
| | | set_all_random_seed(kwargs.get("seed", 0)) |
| | | |
| | | |
| | | device = kwargs.get("device", "cuda") |
| | | if not torch.cuda.is_available() or kwargs.get("ngpu", 1) == 0: |
| | | device = "cpu" |
| | |
| | | vocab_size = len(tokenizer.token_list) |
| | | else: |
| | | vocab_size = -1 |
| | | pdb.set_trace() |
| | | # build frontend |
| | | frontend = kwargs.get("frontend", None) |
| | | |
| | | if frontend is not None: |
| | | frontend_class = tables.frontend_classes.get(frontend) |
| | | frontend = frontend_class(**kwargs["frontend_conf"]) |
| | | kwargs["frontend"] = frontend |
| | | kwargs["input_size"] = frontend.output_size() |
| | | pdb.set_trace() |
| | | |
| | | # build model |
| | | model_class = tables.model_classes.get(kwargs["model"]) |
| | | model = model_class(**kwargs, **kwargs["model_conf"], vocab_size=vocab_size) |
| | | |
| | | model.to(device) |
| | | |
| | | # init_param |
| | |
| | | |
| | | time1 = time.perf_counter() |
| | | with torch.no_grad(): |
| | | pdb.set_trace() |
| | | results, meta_data = model.inference(**batch, **kwargs) |
| | | time2 = time.perf_counter() |
| | | |
| | | pdb.set_trace() |
| | | asr_result_list.extend(results) |
| | | |
| | | # batch_data_time = time_per_frame_s * data_batch_i["speech_lengths"].sum().item() |