From 31e2eb39ad3965931f9df22fce86c708f4d9da95 Mon Sep 17 00:00:00 2001
From: 语帆 <yf352572@alibaba-inc.com>
Date: 星期三, 28 二月 2024 16:14:57 +0800
Subject: [PATCH] test
---
funasr/models/lcbnet/model.py | 7 ++++---
funasr/utils/load_utils.py | 1 -
2 files changed, 4 insertions(+), 4 deletions(-)
diff --git a/funasr/models/lcbnet/model.py b/funasr/models/lcbnet/model.py
index 3b8f3c9..f4caee8 100644
--- a/funasr/models/lcbnet/model.py
+++ b/funasr/models/lcbnet/model.py
@@ -422,7 +422,6 @@
else:
# extract fbank feats
time1 = time.perf_counter()
- pdb.set_trace()
sample_list = load_audio_text_image_video(data_in, fs=frontend.fs, audio_fs=kwargs.get("fs", 16000),
data_type=kwargs.get("data_type", "sound"),
tokenizer=tokenizer)
@@ -443,9 +442,11 @@
encoder_out, encoder_out_lens = self.encode(speech, speech_lengths)
if isinstance(encoder_out, tuple):
encoder_out = encoder_out[0]
- pdb.set_trace()
- ocr = ocr_sample_list[0]
+
+ ocr_list_new = [[x + 1 if x != 0 else x for x in sublist] for sublist in ocr_sample_list]
+ ocr = torch.tensor(ocr_list_new)
ocr_lengths = ocr.new_full([1], dtype=torch.long, fill_value=ocr.size(1))
+ pdb.set_trace()
ocr, ocr_lens, _ = self.text_encoder(ocr, ocr_lengths)
pdb.set_trace()
# c. Passed the encoder result and the beam search
diff --git a/funasr/utils/load_utils.py b/funasr/utils/load_utils.py
index 8b75cbd..87412bd 100644
--- a/funasr/utils/load_utils.py
+++ b/funasr/utils/load_utils.py
@@ -31,7 +31,6 @@
return [load_audio_text_image_video(audio, fs=fs, audio_fs=audio_fs, data_type=data_type, **kwargs) for audio in data_or_path_or_list]
if isinstance(data_or_path_or_list, str) and data_or_path_or_list.startswith('http'): # download url to local file
data_or_path_or_list = download_from_url(data_or_path_or_list)
- pdb.set_trace()
if isinstance(data_or_path_or_list, str) and os.path.exists(data_or_path_or_list): # local file
if data_type is None or data_type == "sound":
data_or_path_or_list, audio_fs = torchaudio.load(data_or_path_or_list)
--
Gitblit v1.9.1