From 40f1f529b0ebc7a571af6be3ca9df67cd5480d66 Mon Sep 17 00:00:00 2001 From: yhliang <429259365@qq.com> Date: 星期四, 20 四月 2023 16:48:26 +0800 Subject: [PATCH] add m2met2 registration form --- docs_m2met2/Track_setting_and_evaluation.md | 4 ++-- 1 files changed, 2 insertions(+), 2 deletions(-) diff --git a/docs_m2met2/Track_setting_and_evaluation.md b/docs_m2met2/Track_setting_and_evaluation.md index b90c17a..2b75fca 100644 --- a/docs_m2met2/Track_setting_and_evaluation.md +++ b/docs_m2met2/Track_setting_and_evaluation.md @@ -1,6 +1,6 @@ # Track & Evaluation -## Speaker-Attributed ASR (Main Track) -The speaker-attributed ASR task poses a unique challenge of transcribing speech from multiple speakers and assigning a speaker label to the transcription. Figure 2 illustrates the difference between the speaker-attributed ASR task and the multi-speaker ASR task. This track allows for the use of the AliMeeting, Aishell4, and Cn-Celeb datasets as constrained data sources during both training and evaluation. The AliMeeting dataset, which was used in the M2MeT challenge, includes Train, Eval, and Test sets. Additionally, a new Test-2023 set, consisting of approximately 10 hours of meeting data recorded in an identical acoustic setting as the AliMeeting corpus, will be released soon for challenge scoring and ranking. It's worth noting that the organizers will not provide the near-field audio, transcriptions, or oracle timestamps. Instead, segments containing multiple speakers will be provided on the Test-2023 set, which can be obtained using a simple voice activity detection (VAD) model. +## Speaker-Attributed ASR +The speaker-attributed ASR task poses a unique challenge of transcribing speech from multiple speakers and assigning a speaker label to the transcription. Figure 2 illustrates the difference between the speaker-attributed ASR task and the multi-speaker ASR task. This track allows for the use of the AliMeeting, Aishell4, and Cn-Celeb datasets as constrained data sources during both training and evaluation. The AliMeeting dataset, which was used in the M2MeT challenge, includes Train, Eval, and Test sets. Additionally, a new Test-2023 set, consisting of approximately 10 hours of meeting data recorded in an identical acoustic setting as the AliMeeting corpus, will be released soon for challenge scoring and ranking. It's worth noting that the organizers will not provide the near-field audio, transcriptions, or oracle timestamps of the Test-2023 set. Instead, segments containing multiple speakers will be provided, which can be obtained using a simple voice activity detection (VAD) model.  -- Gitblit v1.9.1