| .gitignore | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo.sh | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo2.sh | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo2_tmp.sh | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo_pdb.sh | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo_pdb2.sh | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| examples/industrial_data_pretraining/lcbnet/demo_tmp1.sh | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 | |
| funasr/utils/load_utils.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 |
.gitignore
@@ -26,4 +26,5 @@ GPT-SoVITS* examples/*/*/outputs examples/*/*/exp examples/*/*/tmp cmd_read examples/industrial_data_pretraining/lcbnet/demo.py
@@ -10,7 +10,7 @@ # example1 res = model.generate(input='["~/.cache/modelscope/hub/iic/LCB-NET/example/asr_example.wav","~/.cache/modelscope/hub/iic/LCB-NET/example/ocr.txt"]',data_type='["sound", "text"]') res = model.generate(input=("https://www.modelscope.cn/api/v1/models/iic/LCB-NET/repo?Revision=master&FilePath=example/asr_example.wav","https://www.modelscope.cn/api/v1/models/iic/LCB-NET/repo?Revision=master&FilePath=example/ocr.txt"),data_type=("sound", "text")) print(res) examples/industrial_data_pretraining/lcbnet/demo.sh
@@ -1,5 +1,5 @@ file_dir="/nfs/yufan.yf/workspace/github/FunASR/examples/industrial_data_pretraining/lcbnet/exp/speech_lcbnet_contextual_asr-en-16k-bpe-vocab5002-pytorch" CUDA_VISIBLE_DEVICES="0,1,2,3,4,5,6,7" file_dir="/home/yf352572/.cache/modelscope/hub/iic/LCB-NET/" CUDA_VISIBLE_DEVICES="0,1" inference_device="cuda" if [ ${inference_device} == "cuda" ]; then @@ -12,7 +12,7 @@ done fi inference_dir="outputs/slidespeech_dev_beamsearch_new" inference_dir="outputs/slidespeech_dev" _logdir="${inference_dir}/logdir" echo "inference_dir: ${inference_dir}" @@ -39,11 +39,11 @@ python -m funasr.bin.inference \ --config-path=${file_dir} \ --config-name="config.yaml" \ ++init_param=${file_dir}/model.pb \ ++init_param=${file_dir}/model.pt \ ++tokenizer_conf.token_list=${file_dir}/tokens.txt \ ++input=[${_logdir}/wav.${JOB}.scp,${_logdir}/ocr.${JOB}.txt] \ +data_type='["kaldi_ark", "text"]' \ ++tokenizer_conf.bpemodel=${file_dir}/bpe.model \ ++tokenizer_conf.bpemodel=${file_dir}/bpe.pt \ ++output_dir="${inference_dir}/${JOB}" \ ++device="${inference_device}" \ ++ncpu=1 \ examples/industrial_data_pretraining/lcbnet/demo2.sh
File was deleted examples/industrial_data_pretraining/lcbnet/demo2_tmp.sh
File was deleted examples/industrial_data_pretraining/lcbnet/demo_pdb.sh
File was deleted examples/industrial_data_pretraining/lcbnet/demo_pdb2.sh
File was deleted examples/industrial_data_pretraining/lcbnet/demo_tmp1.sh
File was deleted funasr/utils/load_utils.py
@@ -89,8 +89,6 @@ return array def extract_fbank(data, data_len = None, data_type: str="sound", frontend=None, **kwargs): # import pdb; # pdb.set_trace() if isinstance(data, np.ndarray): data = torch.from_numpy(data) if len(data.shape) < 2: