From 4ace5a95b052d338947fc88809a440ccd55cf6b4 Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期四, 16 十一月 2023 16:39:52 +0800
Subject: [PATCH] funasr pages

---
 egs/aishell2/data2vec_pretrain/conf/train_pretrain_transformer.yaml |   70 +++++++++++++++++++---------------
 1 files changed, 39 insertions(+), 31 deletions(-)

diff --git a/egs/aishell2/data2vec_pretrain/conf/train_pretrain_transformer.yaml b/egs/aishell2/data2vec_pretrain/conf/train_pretrain_transformer.yaml
index 4052774..767d8ba 100644
--- a/egs/aishell2/data2vec_pretrain/conf/train_pretrain_transformer.yaml
+++ b/egs/aishell2/data2vec_pretrain/conf/train_pretrain_transformer.yaml
@@ -2,47 +2,52 @@
 # encoder related
 encoder: data2vec_encoder
 encoder_conf:
-  extractor_mode: layer_norm
-  encoder_layerdrop: 0.05
-  dropout_input: 0.0
-  dropout_features: 0.0
-  feature_grad_mult: 1.0
-  encoder_embed_dim: 768
+    extractor_mode: layer_norm
+    encoder_layerdrop: 0.05
+    dropout_input: 0.0
+    dropout_features: 0.0
+    feature_grad_mult: 1.0
+    encoder_embed_dim: 768
 
-  mask_prob: 0.65
-  mask_length: 10
+    mask_prob: 0.65
+    mask_length: 10
 
-  loss_beta: 0
-  loss_scale: null
+    loss_beta: 0
+    loss_scale: null
 
-  instance_norm_target_layer: true
-  average_top_k_layers: 8
+    instance_norm_target_layer: true
+    average_top_k_layers: 8
 
-  pos_conv_depth: 5
-  conv_pos: 95
+    pos_conv_depth: 5
+    conv_pos: 95
 
-  ema_decay: 0.999
-  ema_end_decay: 0.9999
-  ema_anneal_end_step: 30000
-  ema_transformer_only: true
-  ema_layers_only: true
+    ema_decay: 0.999
+    ema_end_decay: 0.9999
+    ema_anneal_end_step: 30000
+    ema_transformer_only: true
+    ema_layers_only: true
 
-  require_same_masks: true
-  mask_dropout: 0
+    require_same_masks: true
+    mask_dropout: 0
 
-log_interval: 50
-normalize: None
+# frontend related
+frontend: wav_frontend
+frontend_conf:
+    fs: 16000
+    window: hamming
+    n_mels: 80
+    frame_length: 25
+    frame_shift: 10
+    lfr_m: 1
+    lfr_n: 1
 
-# minibatch related
-batch_type: length
-batch_bins: 64000
-num_workers: 16
+model: data2vec
 
 # optimization related
 accum_grad: 1
 grad_clip: 5
 patience: none
-max_epoch: 600
+max_epoch: 1800
 val_scheduler_criterion:
     - valid
     - acc
@@ -67,8 +72,8 @@
 # for dataset
 dataset_conf:
     batch_mode: clipping
-    data_names: speech,none
-    data_types: kaldi_ark,none
+    data_names: speech
+    data_types: sound
     shuffle: true
     shuffle_conf:
         shuffle_size: 12800
@@ -76,4 +81,7 @@
     batch_conf:
         batch_type: token
         batch_size: 64000
-    num_workers: 8
\ No newline at end of file
+    num_workers: 8
+
+log_interval: 50
+normalize: None
\ No newline at end of file

--
Gitblit v1.9.1