From 3a4281f4959534b1bf5d01acf0085f4f8e6f2ec8 Mon Sep 17 00:00:00 2001
From: wuhongsheng <664116298@qq.com>
Date: 星期五, 05 七月 2024 00:55:32 +0800
Subject: [PATCH] 优化speakid和语句匹配逻辑,部分解决speakid不从0递增问题 (#1870)
---
fun_text_processing/text_normalization/zh/verbalizers/char.py | 10 ++++------
1 files changed, 4 insertions(+), 6 deletions(-)
diff --git a/fun_text_processing/text_normalization/zh/verbalizers/char.py b/fun_text_processing/text_normalization/zh/verbalizers/char.py
index 4d693fc..b04e441 100644
--- a/fun_text_processing/text_normalization/zh/verbalizers/char.py
+++ b/fun_text_processing/text_normalization/zh/verbalizers/char.py
@@ -1,16 +1,14 @@
-
-
from fun_text_processing.text_normalization.zh.graph_utils import FUN_NOT_QUOTE, GraphFst
from pynini.lib import pynutil
class Char(GraphFst):
- '''
- tokens { char: "浣�" } -> 浣�
- '''
+ """
+ tokens { char: "浣�" } -> 浣�
+ """
def __init__(self, deterministic: bool = True, lm: bool = False):
super().__init__(name="char", kind="verbalize", deterministic=deterministic)
- graph = pynutil.delete("name: \"") + FUN_NOT_QUOTE + pynutil.delete("\"")
+ graph = pynutil.delete('name: "') + FUN_NOT_QUOTE + pynutil.delete('"')
self.fst = graph.optimize()
--
Gitblit v1.9.1