From 0a4e01bd7d789504cc5986fa848e5822bef4dfc9 Mon Sep 17 00:00:00 2001
From: 语帆 <yf352572@alibaba-inc.com>
Date: 星期三, 28 二月 2024 17:18:23 +0800
Subject: [PATCH] atsr

---
 funasr/auto/auto_model.py |    9 +++------
 1 files changed, 3 insertions(+), 6 deletions(-)

diff --git a/funasr/auto/auto_model.py b/funasr/auto/auto_model.py
index e6e08b8..3f99e4d 100644
--- a/funasr/auto/auto_model.py
+++ b/funasr/auto/auto_model.py
@@ -141,7 +141,7 @@
             kwargs = download_model(**kwargs)
         
         set_all_random_seed(kwargs.get("seed", 0))
-        
+
         device = kwargs.get("device", "cuda")
         if not torch.cuda.is_available() or kwargs.get("ngpu", 1) == 0:
             device = "cpu"
@@ -161,19 +161,18 @@
             vocab_size = len(tokenizer.token_list)
         else:
             vocab_size = -1
-        pdb.set_trace()
         # build frontend
         frontend = kwargs.get("frontend", None)
+
         if frontend is not None:
             frontend_class = tables.frontend_classes.get(frontend)
             frontend = frontend_class(**kwargs["frontend_conf"])
             kwargs["frontend"] = frontend
             kwargs["input_size"] = frontend.output_size()
-        pdb.set_trace()
+
         # build model
         model_class = tables.model_classes.get(kwargs["model"])
         model = model_class(**kwargs, **kwargs["model_conf"], vocab_size=vocab_size)
-        
         model.to(device)
         
         # init_param
@@ -235,11 +234,9 @@
         
             time1 = time.perf_counter()
             with torch.no_grad():
-                pdb.set_trace()
                 results, meta_data = model.inference(**batch, **kwargs)
             time2 = time.perf_counter()
             
-            pdb.set_trace()
             asr_result_list.extend(results)
 
             # batch_data_time = time_per_frame_s * data_batch_i["speech_lengths"].sum().item()

--
Gitblit v1.9.1