| funasr/datasets/large_datasets/utils/tokenize.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 |
funasr/datasets/large_datasets/utils/tokenize.py
@@ -37,7 +37,7 @@ vad = -2 if bpe_tokenizer is not None: text = bpe_tokenizer.text2tokens(text) text = bpe_tokenizer.text2tokens("".join(text)) if seg_dict is not None: assert isinstance(seg_dict, dict)