Yu Cao
2025-10-01 c4ac64fd5d24bb3fc8ccc441d36a07c83c8b9015
funasr/tokenizer/sentencepiece_tokenizer.py
@@ -49,3 +49,9 @@
    def get_vocab_size(self):
        return self.sp.GetPieceSize()
    def ids2tokens(self, *args, **kwargs):
        return self.decode(*args, **kwargs)
    def tokens2ids(self, *args, **kwargs):
        return self.encode(*args, **kwargs)