语帆
2024-03-04 1a6d9d5cc422dcd1e6dd5b9c67047d63bc6cd667
atsr
4个文件已修改
5个文件已删除
261 ■■■■■ 已修改文件
.gitignore 1 ●●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo.py 2 ●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo.sh 10 ●●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo2.sh 71 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo2_tmp.sh 71 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo_pdb.sh 18 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo_pdb2.sh 15 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
examples/industrial_data_pretraining/lcbnet/demo_tmp1.sh 71 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
funasr/utils/load_utils.py 2 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
.gitignore
@@ -26,4 +26,5 @@
GPT-SoVITS*
examples/*/*/outputs
examples/*/*/exp
examples/*/*/tmp
cmd_read
examples/industrial_data_pretraining/lcbnet/demo.py
@@ -10,7 +10,7 @@
# example1
res = model.generate(input='["~/.cache/modelscope/hub/iic/LCB-NET/example/asr_example.wav","~/.cache/modelscope/hub/iic/LCB-NET/example/ocr.txt"]',data_type='["sound", "text"]')
res = model.generate(input=("https://www.modelscope.cn/api/v1/models/iic/LCB-NET/repo?Revision=master&FilePath=example/asr_example.wav","https://www.modelscope.cn/api/v1/models/iic/LCB-NET/repo?Revision=master&FilePath=example/ocr.txt"),data_type=("sound", "text"))
print(res)
examples/industrial_data_pretraining/lcbnet/demo.sh
@@ -1,5 +1,5 @@
file_dir="/nfs/yufan.yf/workspace/github/FunASR/examples/industrial_data_pretraining/lcbnet/exp/speech_lcbnet_contextual_asr-en-16k-bpe-vocab5002-pytorch"
CUDA_VISIBLE_DEVICES="0,1,2,3,4,5,6,7"
file_dir="/home/yf352572/.cache/modelscope/hub/iic/LCB-NET/"
CUDA_VISIBLE_DEVICES="0,1"
inference_device="cuda"
if [ ${inference_device} == "cuda" ]; then
@@ -12,7 +12,7 @@
    done
fi
inference_dir="outputs/slidespeech_dev_beamsearch_new"
inference_dir="outputs/slidespeech_dev"
_logdir="${inference_dir}/logdir"
echo "inference_dir: ${inference_dir}"
@@ -39,11 +39,11 @@
        python -m funasr.bin.inference \
        --config-path=${file_dir} \
        --config-name="config.yaml" \
        ++init_param=${file_dir}/model.pb \
        ++init_param=${file_dir}/model.pt \
        ++tokenizer_conf.token_list=${file_dir}/tokens.txt \
        ++input=[${_logdir}/wav.${JOB}.scp,${_logdir}/ocr.${JOB}.txt] \
        +data_type='["kaldi_ark", "text"]' \
        ++tokenizer_conf.bpemodel=${file_dir}/bpe.model \
        ++tokenizer_conf.bpemodel=${file_dir}/bpe.pt \
        ++output_dir="${inference_dir}/${JOB}" \
        ++device="${inference_device}" \
        ++ncpu=1 \
examples/industrial_data_pretraining/lcbnet/demo2.sh
File was deleted
examples/industrial_data_pretraining/lcbnet/demo2_tmp.sh
File was deleted
examples/industrial_data_pretraining/lcbnet/demo_pdb.sh
File was deleted
examples/industrial_data_pretraining/lcbnet/demo_pdb2.sh
File was deleted
examples/industrial_data_pretraining/lcbnet/demo_tmp1.sh
File was deleted
funasr/utils/load_utils.py
@@ -89,8 +89,6 @@
    return array
def extract_fbank(data, data_len = None, data_type: str="sound", frontend=None, **kwargs):
    # import pdb;
    # pdb.set_trace()
    if isinstance(data, np.ndarray):
        data = torch.from_numpy(data)
        if len(data.shape) < 2: