| funasr/tokenizer/char_tokenizer.py | ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史 |
funasr/tokenizer/char_tokenizer.py
@@ -36,6 +36,7 @@ self.remove_non_linguistic_symbols = remove_non_linguistic_symbols self.split_with_space = split_with_space self.seg_dict = None seg_dict = seg_dict if seg_dict is not None else kwargs.get("seg_dict_file", None) if seg_dict is not None: self.seg_dict = load_seg_dict(seg_dict)