雾聪
2024-01-08 f6daebd88eb8aa3d37dc9a063043becb2a7bc8d6
funasr/datasets/small_datasets/preprocessor.py
@@ -361,7 +361,6 @@
                    tokens = seg_tokenize(tokens, self.seg_dict)
            else:
                tokens = self.tokenizer.text2tokens(text)
            text_ints = self.token_id_converter.tokens2ids(tokens)
            data[self.text_name] = np.array(text_ints, dtype=np.int64)
        return data