雾聪
2024-01-03 fcaf910ceb4e07a13bf2d133f46df684b069b3f0
funasr/datasets/small_datasets/preprocessor.py
@@ -361,7 +361,6 @@
                    tokens = seg_tokenize(tokens, self.seg_dict)
            else:
                tokens = self.tokenizer.text2tokens(text)
            text_ints = self.token_id_converter.tokens2ids(tokens)
            data[self.text_name] = np.array(text_ints, dtype=np.int64)
        return data