From ad9b406821dbdd9e4922f90a26ce1712de6cd65a Mon Sep 17 00:00:00 2001
From: hnluo <haoneng.lhn@alibaba-inc.com>
Date: 星期五, 21 七月 2023 14:37:31 +0800
Subject: [PATCH] Update RESULTS.md

---
 egs_modelscope/asr/paraformer/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/RESULTS.md |   55 +++++++++++++++++++++++++++++++++++++++++++++++--------
 1 files changed, 47 insertions(+), 8 deletions(-)

diff --git a/egs_modelscope/asr/paraformer/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/RESULTS.md b/egs_modelscope/asr/paraformer/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/RESULTS.md
index 4e06daf..5f1abb6 100644
--- a/egs_modelscope/asr/paraformer/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/RESULTS.md
+++ b/egs_modelscope/asr/paraformer/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/RESULTS.md
@@ -2,14 +2,6 @@
 - Model link: <https://www.modelscope.cn/models/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/summary>
 - Model size: 220M
 
-# Environments
-- date: `Tue Nov 22 18:48:39 CST 2022`
-- python version: `3.7.12`
-- FunASR version: `0.1.0`
-- pytorch version: `pytorch 1.7.0`
-- Git hash: ``
-- Commit date: ``
-
 # Beachmark Results
 
 ## AISHELL-1
@@ -73,3 +65,50 @@
 |SPEECHIO_ASR_ZH000013| 2.57 | 2.25 |
 |SPEECHIO_ASR_ZH000014| 3.86 | 3.08 |
 |SPEECHIO_ASR_ZH000015| 3.34 | 2.67 |
+
+
+# Fine-tuning Results
+
+## Fine-tuning
+- Train config: 
+  - Training data: aishell-1
+  - Training info: lr 0.0002, batch size 2000, 2 gpu, acc_grad 1, 20 epochs
+  - Decoding info: beam_size 1, average_num 10
+
+| model    | dev cer(%) | test cer(%) |
+|:---------:|:-------------:|:-------------:|
+| Pretrain       | 1.75          |1.95           |
+| Finetune      | 1.62          |1.78           |
+
+- Train config: 
+  - Training data: 16k sichuan dialect
+  - Training info: lr 0.0002, batch size 2000, 2 gpu, acc_grad 1, 20 epochs
+  - Decoding info: beam_size 1, average_num 10
+  
+  
+|   model  | Training Data(h) | cn cer(%) | sichuan cer(%) |
+|:--------:|:-------------:|:-------:|:------------:|
+| Pretrain |               |   8.57  |     19.81    |
+| Finetune |      50      |   8.8   |      12      |
+|          |      100     |   9.24  |     11.63    |
+|          |      200     |   9.82  |     10.47    |
+|          |      300     |   9.95  |     10.44    |
+|          |     1000     |   9.99  |     9.78     |
+
+
+## Lora Fine-tuning
+- Train config: 
+  - Training data: 16k sichuan dialect
+  - Training info: lr 0.0002, batch size 2000, 2 gpu, acc_grad 1, 20 epochs
+  - Lora info: lora_bias: "all", lora_list ['q','v'], lora_rank:8, lora_alpha:16, lora_dropout:0.1
+  - Decoding info: beam_size 1, average_num 10
+  
+|     model     | Training Data(h) | Trainable Parameters(M) | cn cer(%) | sichuan cer(%) |
+|:-------------:|:----------------:|:-----------------------:|:---------:|:--------------:|
+|    Pretrain   |                  |                         |    8.57   |      19.81     |
+|               |                  |                         |           |                |
+|    Finetune   |        50        |          220.9          |    8.8    |       12       |
+| Lora Finetune |        50        |           2.29          |    9.13   |      12.13     |
+|               |                  |                         |           |                |
+|    Finetune   |        200       |          220.9          |    9.82   |      10.47     |
+| Lora Finetune |        200       |           2.29          |    9.21   |      11.28     |

--
Gitblit v1.9.1