shixian.shi
2024-01-10 2d0c8274a6dd0db9d5f7a00e71401bf7b4d65553
funasr/tokenizer/phoneme_tokenizer.py
@@ -13,7 +13,7 @@
from funasr.tokenizer.abs_tokenizer import AbsTokenizer
g2p_choices = [
g2p_classes = [
    None,
    "g2p_en",
    "g2p_en_no_space",
@@ -363,6 +363,7 @@
        non_linguistic_symbols: Union[Path, str, Iterable[str]] = None,
        space_symbol: str = "<space>",
        remove_non_linguistic_symbols: bool = False,
        **kwargs,
    ):
        if g2p_type is None:
            self.g2p = split_by_space