funasr/tokenizer/word_tokenizer.py @@ -14,6 +14,7 @@ delimiter: str = None, non_linguistic_symbols: Union[Path, str, Iterable[str]] = None, remove_non_linguistic_symbols: bool = False, **kwargs, ): self.delimiter = delimiter