From 08ee9e6aacc2e306211d393f6e8ce3a7f3620102 Mon Sep 17 00:00:00 2001 From: yhliang <68215459+yhliang-aslp@users.noreply.github.com> Date: 星期四, 10 八月 2023 20:46:21 +0800 Subject: [PATCH] Add modular SA-ASR recipe for M2MeT2.0 (#831) --- docs/reference/papers.md | 4 +++- 1 files changed, 3 insertions(+), 1 deletions(-) diff --git a/docs/reference/papers.md b/docs/reference/papers.md index 33bf72f..22da0db 100644 --- a/docs/reference/papers.md +++ b/docs/reference/papers.md @@ -3,7 +3,9 @@ FunASR have implemented the following paper code ### Speech Recognition -- [Paraformer: Fast and Accurate Parallel Transformer for Non-autoregressive End-to-End Speech Recognition](https://arxiv.org/abs/2206.08317), INTERSPEECH 2022. +- [FunASR: A Fundamental End-to-End Speech Recognition Toolkit](https://arxiv.org/abs/2305.11013), INTERSPEECH 2023 +- [BAT: Boundary aware transducer for memory-efficient and low-latency ASR](https://arxiv.org/abs/2305.11571), INTERSPEECH 2023 +- [Paraformer: Fast and Accurate Parallel Transformer for Non-autoregressive End-to-End Speech Recognition](https://arxiv.org/abs/2206.08317), INTERSPEECH 2022 - [Universal ASR: Unifying Streaming and Non-Streaming ASR Using a Single Encoder-Decoder Model](https://arxiv.org/abs/2010.14099), arXiv preprint arXiv:2010.14099, 2020. - [San-m: Memory equipped self-attention for end-to-end speech recognition](https://arxiv.org/pdf/2006.01713), INTERSPEECH 2020 - [Streaming Chunk-Aware Multihead Attention for Online End-to-End Speech Recognition](https://arxiv.org/abs/2006.01712), INTERSPEECH 2020 -- Gitblit v1.9.1