zhu-gu-an
2024-01-13 49e8e9d8fc1209c347aa2c2c65c6eb067b9f79d4
funasr/tokenizer/word_tokenizer.py
@@ -14,7 +14,6 @@
        delimiter: str = None,
        non_linguistic_symbols: Union[Path, str, Iterable[str]] = None,
        remove_non_linguistic_symbols: bool = False,
        **kwargs,
    ):
        self.delimiter = delimiter