语帆
2024-02-28 0a4e01bd7d789504cc5986fa848e5822bef4dfc9
funasr/auto/auto_model.py
@@ -141,7 +141,7 @@
            kwargs = download_model(**kwargs)
        
        set_all_random_seed(kwargs.get("seed", 0))
        device = kwargs.get("device", "cuda")
        if not torch.cuda.is_available() or kwargs.get("ngpu", 1) == 0:
            device = "cpu"
@@ -172,14 +172,11 @@
        # build model
        model_class = tables.model_classes.get(kwargs["model"])
        pdb.set_trace()
        model = model_class(**kwargs, **kwargs["model_conf"], vocab_size=vocab_size)
        pdb.set_trace()
        model.to(device)
        
        # init_param
        init_param = kwargs.get("init_param", None)
        pdb.set_trace()
        if init_param is not None:
            logging.info(f"Loading pretrained params from {init_param}")
            load_pretrained_model(
@@ -237,11 +234,9 @@
        
            time1 = time.perf_counter()
            with torch.no_grad():
                pdb.set_trace()
                results, meta_data = model.inference(**batch, **kwargs)
            time2 = time.perf_counter()
            
            pdb.set_trace()
            asr_result_list.extend(results)
            # batch_data_time = time_per_frame_s * data_batch_i["speech_lengths"].sum().item()