From 0892f5ce5240fde47fdcfc6f4faea8bfad6dc0ce Mon Sep 17 00:00:00 2001
From: speech_asr <wangjiaming.wjm@alibaba-inc.com>
Date: 星期三, 08 三月 2023 16:49:06 +0800
Subject: [PATCH] update eend_ola

---
 funasr/modules/eend_ola/utils/losses.py |   12 +-----------
 1 files changed, 1 insertions(+), 11 deletions(-)

diff --git a/funasr/modules/eend_ola/utils/losses.py b/funasr/modules/eend_ola/utils/losses.py
index 97443bc..af0181d 100644
--- a/funasr/modules/eend_ola/utils/losses.py
+++ b/funasr/modules/eend_ola/utils/losses.py
@@ -8,19 +8,9 @@
 def standard_loss(ys, ts, label_delay=0):
     losses = [F.binary_cross_entropy(torch.sigmoid(y), t) * len(y) for y, t in zip(ys, ts)]
     loss = torch.sum(torch.stack(losses))
-    n_frames = torch.from_numpy(np.array(np.sum([t.shape[0] for t in ts]))).to(torch.float32).to(ys[0].device)  # 璁$畻鎬荤殑甯ф暟
+    n_frames = torch.from_numpy(np.array(np.sum([t.shape[0] for t in ts]))).to(torch.float32).to(ys[0].device)
     loss = loss / n_frames
     return loss
-
-
-def batch_pit_loss(ys, ts, label_delay=0):
-    loss_w_labels = [pit_loss(y, t)
-                     for (y, t) in zip(ys, ts)]
-    losses, labels = zip(*loss_w_labels)
-    loss = torch.sum(torch.stack(losses))
-    n_frames = torch.sum(torch.stack([t.shape[0] for t in ts]))
-    loss = loss / n_frames
-    return loss, labels
 
 
 def batch_pit_n_speaker_loss(ys, ts, n_speakers_list):

--
Gitblit v1.9.1