游雁
2024-04-29 2779602177ae5374547c7a7e17de0b11a166326d
funasr/tokenizer/sentencepiece_tokenizer.py
@@ -8,11 +8,10 @@
from funasr.tokenizer.abs_tokenizer import BaseTokenizer
from funasr.register import tables
@tables.register("tokenizer_classes", "SentencepiecesTokenizer")
class SentencepiecesTokenizer(BaseTokenizer):
    def __init__(self, bpemodel: Union[Path, str],
                 **kwargs
                 ):
    def __init__(self, bpemodel: Union[Path, str], **kwargs):
        super().__init__(**kwargs)
        self.bpemodel = str(bpemodel)
        # NOTE(kamo):
@@ -45,4 +44,4 @@
    def decode(self, line: List[int]):
        self._build_sentence_piece_processor()
        return self.sp.DecodeIds(line)
        return self.sp.DecodeIds(line)