From d77910eb6d171727f2350e45c31c91436c4c8891 Mon Sep 17 00:00:00 2001
From: 游雁 <zhifu.gzf@alibaba-inc.com>
Date: 星期一, 11 十二月 2023 13:42:40 +0800
Subject: [PATCH] funasr2
---
outputs/2023-12-11/13-19-05/.hydra/config.yaml | 5
outputs/2023-12-11/13-07-13/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/11-44-36/.hydra/overrides.yaml | 5
outputs/2023-12-11/13-07-13/.hydra/overrides.yaml | 5
outputs/2023-12-11/11-33-27/.hydra/overrides.yaml | 5
outputs/2023-12-11/13-39-37/train_cli.log | 6
outputs/2023-12-11/13-24-15/.hydra/hydra.yaml | 156 ++
examples/aishell/conf/train_asr_paraformer_conformer_12e_6d_2048_256.yaml | 113 +
outputs/2023-12-11/11-37-13/.hydra/config.yaml | 5
funasr/cli/trainer.py | 48
outputs/2023-12-11/10-45-48/.hydra/overrides.yaml | 4
outputs/2023-12-11/13-37-51/train_cli.log | 6
outputs/2023-12-11/13-37-51/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/13-24-15/.hydra/config.yaml | 5
outputs/2023-12-11/11-26-16/train_cli.log | 0
outputs/2023-12-11/13-12-46/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/13-16-01/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/11-37-13/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/11-48-01/.hydra/config.yaml | 5
outputs/2023-12-11/11-36-25/train_cli.log | 2
outputs/2023-12-11/11-44-36/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/11-48-01/train_cli.log | 3
funasr/datasets/dataset_jsonl.py | 20
outputs/2023-12-11/11-37-13/.hydra/overrides.yaml | 5
outputs/2023-12-11/13-07-13/train_cli.log | 3
outputs/2023-12-11/10-45-48/.hydra/config.yaml | 4
outputs/2023-12-11/10-39-53/.hydra/hydra.yaml | 154 ++
outputs/2023-12-11/11-32-18/.hydra/config.yaml | 4
outputs/2023-12-11/10-35-57/train_cli.log | 0
outputs/2023-12-11/11-26-16/.hydra/hydra.yaml | 158 ++
outputs/2023-12-11/10-37-01/train_cli.log | 0
outputs/2023-12-11/11-32-18/.hydra/overrides.yaml | 4
outputs/2023-12-11/13-06-39/.hydra/overrides.yaml | 5
outputs/2023-12-11/10-41-29/train_cli.log | 0
outputs/2023-12-11/11-36-25/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/11-36-25/.hydra/overrides.yaml | 5
outputs/2023-12-11/11-26-16/.hydra/overrides.yaml | 4
outputs/2023-12-11/10-35-57/.hydra/hydra.yaml | 154 ++
outputs/2023-12-11/13-07-13/.hydra/config.yaml | 5
outputs/2023-12-11/13-12-46/.hydra/config.yaml | 5
outputs/2023-12-11/10-41-29/.hydra/config.yaml | 74 +
outputs/2023-12-11/10-44-08/train_cli.log | 0
outputs/2023-12-11/10-45-48/train_cli.log | 0
outputs/2023-12-11/13-37-51/.hydra/config.yaml | 5
funasr/utils/download_from_hub.py | 63
outputs/2023-12-11/13-19-05/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/13-06-39/train_cli.log | 0
outputs/2023-12-11/10-39-53/.hydra/config.yaml | 1
funasr/utils/name_maps_from_hub.py | 16
outputs/2023-12-11/11-33-27/.hydra/config.yaml | 5
outputs/2023-12-11/10-39-53/train_cli.log | 0
outputs/2023-12-11/13-19-05/.hydra/overrides.yaml | 5
outputs/2023-12-11/10-39-53/.hydra/overrides.yaml | 1
outputs/2023-12-11/11-33-27/.hydra/hydra.yaml | 156 ++
examples/aishell/run.sh | 9
outputs/2023-12-11/10-37-01/.hydra/overrides.yaml | 1
funasr/cli/models/paraformer.py | 3
outputs/2023-12-11/11-44-36/.hydra/config.yaml | 5
outputs/2023-12-11/10-45-48/.hydra/hydra.yaml | 158 ++
outputs/2023-12-11/10-37-01/.hydra/hydra.yaml | 154 ++
outputs/2023-12-11/13-24-15/.hydra/overrides.yaml | 5
outputs/2023-12-11/10-44-08/.hydra/config.yaml | 4
outputs/2023-12-11/11-36-25/.hydra/config.yaml | 5
outputs/2023-12-11/13-37-51/.hydra/overrides.yaml | 5
outputs/2023-12-11/10-37-01/.hydra/config.yaml | 1
outputs/2023-12-11/10-44-08/.hydra/overrides.yaml | 4
outputs/2023-12-11/11-48-01/.hydra/overrides.yaml | 5
outputs/2023-12-11/11-26-16/.hydra/config.yaml | 4
outputs/2023-12-11/13-24-15/train_cli.log | 6
outputs/2023-12-11/11-48-01/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/13-16-01/.hydra/overrides.yaml | 5
outputs/2023-12-11/10-41-29/.hydra/overrides.yaml | 4
outputs/2023-12-11/13-06-39/.hydra/config.yaml | 5
outputs/2023-12-11/13-06-39/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/13-16-01/train_cli.log | 6
outputs/2023-12-11/11-37-13/train_cli.log | 2
outputs/2023-12-11/10-35-57/.hydra/overrides.yaml | 1
outputs/2023-12-11/13-16-01/.hydra/config.yaml | 5
outputs/2023-12-11/13-19-05/train_cli.log | 6
examples/industrial_data_pretraining/paraformer-large/run.sh | 12
outputs/2023-12-11/10-44-08/.hydra/hydra.yaml | 158 ++
outputs/2023-12-11/13-39-37/.hydra/hydra.yaml | 156 ++
outputs/2023-12-11/10-35-57/.hydra/config.yaml | 1
outputs/2023-12-11/13-12-46/train_cli.log | 5
outputs/2023-12-11/13-39-37/.hydra/config.yaml | 5
outputs/2023-12-11/11-44-36/train_cli.log | 2
outputs/2023-12-11/11-33-27/train_cli.log | 1
funasr/cli/train_cli.py | 32
examples/industrial_data_pretraining/paraformer-large/conf/finetune.yaml | 120 +
outputs/2023-12-11/10-41-29/.hydra/hydra.yaml | 158 ++
outputs/2023-12-11/11-32-18/.hydra/hydra.yaml | 155 ++
outputs/2023-12-11/11-32-18/train_cli.log | 1
outputs/2023-12-11/13-39-37/.hydra/overrides.yaml | 5
outputs/2023-12-11/13-12-46/.hydra/overrides.yaml | 5
94 files changed, 3,986 insertions(+), 22 deletions(-)
diff --git a/examples/aishell/conf/train_asr_paraformer_conformer_12e_6d_2048_256.yaml b/examples/aishell/conf/train_asr_paraformer_conformer_12e_6d_2048_256.yaml
new file mode 100644
index 0000000..94b7f6d
--- /dev/null
+++ b/examples/aishell/conf/train_asr_paraformer_conformer_12e_6d_2048_256.yaml
@@ -0,0 +1,113 @@
+
+# network architecture
+model: funasr.cli.models.paraformer:Paraformer
+model_conf:
+ ctc_weight: 0.3
+ lsm_weight: 0.1
+ length_normalized_loss: false
+ predictor_weight: 1.0
+ sampling_ratio: 0.4
+ use_1st_decoder_loss: true
+
+
+# encoder related
+encoder: conformer
+encoder_conf:
+ output_size: 256 # dimension of attention
+ attention_heads: 4
+ linear_units: 2048 # the number of units of position-wise feed forward
+ num_blocks: 12 # the number of encoder blocks
+ dropout_rate: 0.1
+ positional_dropout_rate: 0.1
+ attention_dropout_rate: 0.0
+ input_layer: conv2d # encoder architecture type
+ normalize_before: true
+ pos_enc_layer_type: rel_pos
+ selfattention_layer_type: rel_selfattn
+ activation_type: swish
+ macaron_style: true
+ use_cnn_module: true
+ cnn_module_kernel: 15
+
+# decoder related
+decoder: paraformer_decoder_san
+decoder_conf:
+ attention_heads: 4
+ linear_units: 2048
+ num_blocks: 6
+ dropout_rate: 0.1
+ positional_dropout_rate: 0.1
+ self_attention_dropout_rate: 0.0
+ src_attention_dropout_rate: 0.0
+
+# frontend related
+frontend: wav_frontend
+frontend_conf:
+ fs: 16000
+ window: hamming
+ n_mels: 80
+ frame_length: 25
+ frame_shift: 10
+ lfr_m: 1
+ lfr_n: 1
+
+
+train_conf:
+ accum_grad: 1
+ grad_clip: 5
+ max_epoch: 150
+ val_scheduler_criterion:
+ - valid
+ - acc
+ best_model_criterion:
+ - - valid
+ - acc
+ - max
+ keep_nbest_models: 10
+ log_interval: 50
+
+optim: adam
+optim_conf:
+ lr: 0.0005
+scheduler: warmuplr
+scheduler_conf:
+ warmup_steps: 30000
+
+specaug: specaug
+specaug_conf:
+ apply_time_warp: true
+ time_warp_window: 5
+ time_warp_mode: bicubic
+ apply_freq_mask: true
+ freq_mask_width_range:
+ - 0
+ - 30
+ num_freq_mask: 2
+ apply_time_mask: true
+ time_mask_width_range:
+ - 0
+ - 40
+ num_time_mask: 2
+
+predictor: cif_predictor
+predictor_conf:
+ idim: 256
+ threshold: 1.0
+ l_order: 1
+ r_order: 1
+ tail_threshold: 0.45
+
+dataset_conf:
+ data_names: speech,text
+ data_types: sound,text
+ shuffle: True
+ shuffle_conf:
+ shuffle_size: 2048
+ sort_size: 500
+ batch_conf:
+ batch_type: example
+ batch_size: 2
+ num_workers: 8
+
+
+normalize: null
\ No newline at end of file
diff --git a/examples/aishell/run.sh b/examples/aishell/run.sh
new file mode 100644
index 0000000..786691f
--- /dev/null
+++ b/examples/aishell/run.sh
@@ -0,0 +1,9 @@
+
+cmd="funasr_cli/cli/train_cli.py"
+
+python $cmd \
+--config-path "/Users/zhifu/funasr_github/test_local/funasr_cli_egs" \
+--config-name "config.yaml" \
++token_list="/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt" \
++train_data_set_list="/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl" \
++output_dir="/nfs/zhifu.gzf/ckpt/funasr2/exp1"
\ No newline at end of file
diff --git a/examples/industrial_data_pretraining/paraformer-large/conf/finetune.yaml b/examples/industrial_data_pretraining/paraformer-large/conf/finetune.yaml
new file mode 100644
index 0000000..880aad9
--- /dev/null
+++ b/examples/industrial_data_pretraining/paraformer-large/conf/finetune.yaml
@@ -0,0 +1,120 @@
+
+# network architecture
+model: funasr.cli.models.paraformer:Paraformer
+model_conf:
+ ctc_weight: 0.0
+ lsm_weight: 0.1
+ length_normalized_loss: true
+ predictor_weight: 1.0
+ predictor_bias: 1
+ sampling_ratio: 0.75
+
+# encoder
+encoder: sanm
+encoder_conf:
+ output_size: 512
+ attention_heads: 4
+ linear_units: 2048
+ num_blocks: 50
+ dropout_rate: 0.1
+ positional_dropout_rate: 0.1
+ attention_dropout_rate: 0.1
+ input_layer: pe
+ pos_enc_class: SinusoidalPositionEncoder
+ normalize_before: true
+ kernel_size: 11
+ sanm_shfit: 0
+ selfattention_layer_type: sanm
+
+# decoder
+decoder: paraformer_decoder_sanm
+decoder_conf:
+ attention_heads: 4
+ linear_units: 2048
+ num_blocks: 16
+ dropout_rate: 0.1
+ positional_dropout_rate: 0.1
+ self_attention_dropout_rate: 0.1
+ src_attention_dropout_rate: 0.1
+ att_layer_num: 16
+ kernel_size: 11
+ sanm_shfit: 0
+
+predictor: cif_predictor_v2
+predictor_conf:
+ idim: 512
+ threshold: 1.0
+ l_order: 1
+ r_order: 1
+ tail_threshold: 0.45
+
+# frontend related
+frontend: wav_frontend
+frontend_conf:
+ fs: 16000
+ window: hamming
+ n_mels: 80
+ frame_length: 25
+ frame_shift: 10
+ lfr_m: 7
+ lfr_n: 6
+
+specaug: specaug_lfr
+specaug_conf:
+ apply_time_warp: false
+ time_warp_window: 5
+ time_warp_mode: bicubic
+ apply_freq_mask: true
+ freq_mask_width_range:
+ - 0
+ - 30
+ lfr_rate: 6
+ num_freq_mask: 1
+ apply_time_mask: true
+ time_mask_width_range:
+ - 0
+ - 12
+ num_time_mask: 1
+
+train_conf:
+ accum_grad: 1
+ grad_clip: 5
+ max_epoch: 150
+ val_scheduler_criterion:
+ - valid
+ - acc
+ best_model_criterion:
+ - - valid
+ - acc
+ - max
+ keep_nbest_models: 10
+ log_interval: 50
+
+optim: adam
+optim_conf:
+ lr: 0.0005
+scheduler: warmuplr
+scheduler_conf:
+ warmup_steps: 30000
+
+
+dataset_conf:
+ data_names: speech,text
+ data_types: sound,text
+ shuffle: True
+ shuffle_conf:
+ shuffle_size: 2048
+ sort_size: 500
+ batch_conf:
+ batch_type: example
+ batch_size: 2
+ num_workers: 8
+
+split_with_space: true
+input_size: 560
+ctc_conf:
+ dropout_rate: 0.0
+ ctc_type: builtin
+ reduce: true
+ ignore_nan_grad: true
+normalize: null
\ No newline at end of file
diff --git a/examples/industrial_data_pretraining/paraformer-large/run.sh b/examples/industrial_data_pretraining/paraformer-large/run.sh
new file mode 100644
index 0000000..9b40b81
--- /dev/null
+++ b/examples/industrial_data_pretraining/paraformer-large/run.sh
@@ -0,0 +1,12 @@
+
+cmd="funasr/cli/train_cli.py"
+
+python $cmd \
++model_pretrain="/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch" \
++token_list="/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt" \
++train_data_set_list="/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl" \
++output_dir="/Users/zhifu/Downloads/ckpt/funasr2/exp2" \
++device="cpu"
+
+#--config-path "/Users/zhifu/funasr_github/examples/industrial_data_pretraining/paraformer-large/conf" \
+#--config-name "finetune.yaml" \
\ No newline at end of file
diff --git a/funasr/cli/models/paraformer.py b/funasr/cli/models/paraformer.py
index ee8c0b4..7ca80f5 100644
--- a/funasr/cli/models/paraformer.py
+++ b/funasr/cli/models/paraformer.py
@@ -193,6 +193,7 @@
self.decoder.embed = None
self.use_1st_decoder_loss = use_1st_decoder_loss
+ self.length_normalized_loss = length_normalized_loss
def forward(
self,
@@ -302,6 +303,8 @@
stats["loss"] = torch.clone(loss.detach())
# force_gatherable: to-device and to-tensor if scalar for DataParallel
+ if self.length_normalized_loss:
+ batch_size = (text_lengths + self.predictor_bias).sum()
loss, stats, weight = force_gatherable((loss, stats, batch_size), loss.device)
return loss, stats, weight
diff --git a/funasr/cli/train_cli.py b/funasr/cli/train_cli.py
index 54cd2e8..c62153e 100644
--- a/funasr/cli/train_cli.py
+++ b/funasr/cli/train_cli.py
@@ -25,17 +25,25 @@
import torch.distributed as dist
from torch.nn.parallel import DistributedDataParallel as DDP
from torch.distributed.fsdp import FullyShardedDataParallel as FSDP
-
+from funasr.utils.download_from_hub import download_model
def preprocess_config(cfg: DictConfig):
for key, value in cfg.items():
if value == 'None':
cfg[key] = None
+@hydra.main(config_name=None, version_base=None)
+def main_hydra(kwargs: DictConfig):
+ import pdb; pdb.set_trace()
+ if kwargs.get("model_pretrain"):
+ kwargs = download_model(**kwargs)
+
+ import pdb;
+ pdb.set_trace()
+ main(**kwargs)
-@hydra.main()
-def main(kwargs: DictConfig):
+def main(**kwargs):
# preprocess_config(kwargs)
# import pdb; pdb.set_trace()
# set random seed
@@ -76,10 +84,10 @@
# init_param
init_param = kwargs.get("init_param", None)
if init_param is not None:
- init_param = eval(init_param)
+ init_param = init_param
if isinstance(init_param, Sequence):
init_param = (init_param,)
- logging.info("init_param is not None: ", init_param)
+ logging.info("init_param is not None: %s", init_param)
for p in init_param:
logging.info(f"Loading pretrained params from {p}")
load_pretrained_model(
@@ -99,7 +107,7 @@
freeze_param = eval(freeze_param)
if isinstance(freeze_param, Sequence):
freeze_param = (freeze_param,)
- logging.info("freeze_param is not None: ", freeze_param)
+ logging.info("freeze_param is not None: %s", freeze_param)
for t in freeze_param:
for k, p in model.named_parameters():
if k.startswith(t + ".") or k == t:
@@ -138,9 +146,15 @@
dataloader_tr = torch.utils.data.DataLoader(dataset_tr,
collate_fn=dataset_tr.collator,
batch_sampler=batch_sampler,
- num_workers=kwargs.get("num_workers", 0),
+ num_workers=kwargs.get("dataset_conf").get("num_workers", 4),
pin_memory=True)
-
+
+ if (use_ddp or use_fsdp) and dist.get_rank() == 0 or not (use_ddp or use_fsdp) and local_rank == 0:
+ os.makedirs(kwargs.get("output_dir", "./"), exist_ok=True)
+ yaml_file = os.path.join(kwargs.get("output_dir", "./"), "config.yaml")
+ OmegaConf.save(config=kwargs, f=yaml_file)
+ logging.info("config.yaml is saved to: %s", yaml_file)
+
trainer = Trainer(
model=model,
optim=optim,
@@ -160,4 +174,4 @@
if __name__ == "__main__":
- main()
\ No newline at end of file
+ main_hydra()
\ No newline at end of file
diff --git a/funasr/cli/trainer.py b/funasr/cli/trainer.py
index 28a843b..0178767 100644
--- a/funasr/cli/trainer.py
+++ b/funasr/cli/trainer.py
@@ -2,6 +2,7 @@
import os
from funasr.torch_utils.device_funcs import to_device
import logging
+import time
from tqdm import tqdm
from contextlib import nullcontext
import torch.distributed as dist
@@ -53,19 +54,27 @@
self.dataloader_train = dataloader_train
self.dataloader_val = dataloader_val
self.output_dir = kwargs.get('output_dir', './')
- self.resume = kwargs.get('resume', None)
+ self.resume = kwargs.get('resume', True)
self.start_epoch = 1
self.max_epoch = kwargs.get('max_epoch', 100)
self.local_rank = local_rank
- self.rank = dist.get_rank()
- self.world_size = dist.get_world_size()
self.use_ddp = use_ddp
self.use_fsdp = use_fsdp
- self.device = torch.device("cuda", local_rank)
+ self.device = next(model.parameters()).device
self.kwargs = kwargs
if self.resume:
self._resume_checkpoint(self.resume)
+
+ try:
+ rank = dist.get_rank()
+ world_size = dist.get_world_size()
+ except:
+ rank = 0
+ world_size = 1
+ logging.warning("distributed is not initialized, only single shard")
+ self.rank = rank
+ self.world_size = world_size
def _save_checkpoint(self, epoch):
"""
@@ -117,6 +126,7 @@
if dist.get_rank() == 0:
self._save_checkpoint(epoch)
self.scheduler.step()
+ break
def _train_epoch(self, epoch):
"""
@@ -132,13 +142,21 @@
accum_grad = self.kwargs.get("accum_grad", 1)
# Initialize the gradient accumulation
self.optim.zero_grad()
-
+ speed_stats = {}
+ time5 = time.perf_counter()
for batch_idx, batch in enumerate(self.dataloader_train):
+ time1 = time.perf_counter()
+ speed_stats["data_load"] = f"{time1-time5:0.3f}"
+ # import pdb;
+ # pdb.set_trace()
batch = to_device(batch, self.device)
my_context = self.model.no_sync if batch_idx % accum_grad != 0 else nullcontext
with my_context():
+ time2 = time.perf_counter()
retval = self.model(**batch)
+ time3 = time.perf_counter()
+ speed_stats["forward_time"] = f"{time3 - time2:0.3f}"
loss, stats, weight = retval
stats = {k: v for k, v in stats.items() if v is not None}
if self.use_ddp or self.use_fsdp:
@@ -154,6 +172,8 @@
# Scale the loss since we're not updating for every mini-batch
loss = loss / accum_grad
loss.backward()
+ time4 = time.perf_counter()
+ speed_stats["backward_time"] = f"{time4 - time3:0.3f}"
# Perform an optimizer step only after accumulating enough gradients
if (batch_idx + 1) % accum_grad == 0 or (batch_idx + 1) == len(self.dataloader_train):
@@ -176,12 +196,26 @@
self.scheduler.step()
# Clear gradients for the next accumulation stage
self.optim.zero_grad()
+ total_time = f"{time.perf_counter() - time5:0.3f}"
+ time5 = time.perf_counter()
+ speed_stats["optim_time"] = f"{time5 - time4:0.3f}"
+
+ speed_stats["total_time"] = total_time
+
pbar.update(1)
if self.local_rank == 0:
- pbar.set_description(
- f"Training Epoch: {epoch + 1}/{self.max_epoch}, step {batch_idx}/{len(self.dataloader_train)} (loss: {loss.detach().float():.3f}, {[(k, round(v.cpu().item(), 3)) for k, v in stats.items()]})")
+ description = (
+ f"Epoch: {epoch + 1}/{self.max_epoch}, "
+ f"step {batch_idx}/{len(self.dataloader_train)}, "
+ f"{speed_stats}, "
+ f"(loss: {loss.detach().float():.3f}), "
+ f"{[(k, round(v.cpu().item(), 3)) for k, v in stats.items()]}"
+ )
+ pbar.set_description(description)
+ if batch_idx == 2:
+ break
pbar.close()
def _validate_epoch(self, epoch):
diff --git a/funasr/datasets/dataset_jsonl.py b/funasr/datasets/dataset_jsonl.py
index eef67c5..7f2cd83 100644
--- a/funasr/datasets/dataset_jsonl.py
+++ b/funasr/datasets/dataset_jsonl.py
@@ -6,6 +6,7 @@
import librosa
import torchaudio
import time
+import logging
def load_audio(audio_path: str, fs: int=16000):
audio = None
@@ -41,8 +42,7 @@
def __init__(self, path):
super().__init__()
- # data_parallel_size = dist.get_world_size()
- data_parallel_size = 1
+
contents = []
with open(path, encoding='utf-8') as fin:
for line in fin:
@@ -66,12 +66,20 @@
self.contents = []
total_num = len(contents)
- num_per_rank = total_num // data_parallel_size
- # rank = dist.get_rank()
- rank = 0
+ try:
+ rank = dist.get_rank()
+ world_size = dist.get_world_size()
+ except:
+ rank = 0
+ world_size = 1
+ logging.warning("distributed is not initialized, only single shard")
+ num_per_rank = total_num // world_size
+
+ # rank = 0
# import ipdb; ipdb.set_trace()
self.contents = contents[rank * num_per_rank:(rank + 1) * num_per_rank]
-
+
+ logging.info("in rank: {}, num of samplers: {}, total_num of samplers across ranks: {}".format(rank, len(self.contents), len(contents)))
def __len__(self):
return len(self.contents)
diff --git a/funasr/utils/download_from_hub.py b/funasr/utils/download_from_hub.py
new file mode 100644
index 0000000..d6e4ab4
--- /dev/null
+++ b/funasr/utils/download_from_hub.py
@@ -0,0 +1,63 @@
+import os
+from omegaconf import OmegaConf
+import torch
+from funasr.utils.name_maps_from_hub import name_maps_ms, name_maps_hf
+
+def download_model(**kwargs):
+ model_hub = kwargs.get("model_hub", "ms")
+ if model_hub == "ms":
+ kwargs = download_fr_ms(**kwargs)
+
+ return kwargs
+
+def download_fr_ms(**kwargs):
+ model_or_path = kwargs.get("model_pretrain")
+ model_revision = kwargs.get("model_pretrain_revision")
+ if not os.path.exists(model_or_path):
+ model_or_path = get_or_download_model_dir(model_or_path, model_revision, third_party="funasr")
+
+ config = os.path.join(model_or_path, "config.yaml")
+ assert os.path.exists(config), "{} is not exist!".format(config)
+ cfg = OmegaConf.load(config)
+ kwargs = OmegaConf.merge(cfg, kwargs)
+ init_param = os.path.join(model_or_path, "model.pb")
+ kwargs["init_param"] = init_param
+ kwargs["token_list"] = os.path.join(model_or_path, "tokens.txt")
+
+ return kwargs
+
+def get_or_download_model_dir(
+ model,
+ model_revision=None,
+ third_party=None):
+ """ Get local model directory or download model if necessary.
+
+ Args:
+ model (str): model id or path to local model directory.
+ model_revision (str, optional): model version number.
+ third_party (str, optional): in which third party library
+ this function is called.
+ """
+ from modelscope.hub.check_model import check_local_model_is_latest
+ from modelscope.hub.snapshot_download import snapshot_download
+
+ from modelscope.utils.constant import Invoke, ThirdParty
+
+ if os.path.exists(model):
+ model_cache_dir = model if os.path.isdir(
+ model) else os.path.dirname(model)
+ check_local_model_is_latest(
+ model_cache_dir,
+ user_agent={
+ Invoke.KEY: Invoke.LOCAL_TRAINER,
+ ThirdParty.KEY: third_party
+ })
+ else:
+ model_cache_dir = snapshot_download(
+ model,
+ revision=model_revision,
+ user_agent={
+ Invoke.KEY: Invoke.TRAINER,
+ ThirdParty.KEY: third_party
+ })
+ return model_cache_dir
\ No newline at end of file
diff --git a/funasr/utils/name_maps_from_hub.py b/funasr/utils/name_maps_from_hub.py
new file mode 100644
index 0000000..90b44cd
--- /dev/null
+++ b/funasr/utils/name_maps_from_hub.py
@@ -0,0 +1,16 @@
+
+
+name_maps_ms = {
+ "paraformer-zh": "damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch",
+ "paraformer-zh-spk": "damo/speech_paraformer-large-vad-punc-spk_asr_nat-zh-cn",
+ "paraformer-en": "damo/speech_paraformer-large-vad-punc_asr_nat-en-16k-common-vocab10020",
+ "paraformer-en-spk": "damo/speech_paraformer-large-vad-punc_asr_nat-en-16k-common-vocab10020",
+ "paraformer-zh-streaming": "damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-online",
+ "fsmn-vad": "damo/speech_fsmn_vad_zh-cn-16k-common-pytorch",
+ "ct-punc": "damo/punc_ct-transformer_cn-en-common-vocab471067-large",
+ "fa-zh": "damo/speech_timestamp_prediction-v1-16k-offline",
+}
+
+name_maps_hf = {
+
+}
\ No newline at end of file
diff --git a/outputs/2023-12-11/10-35-57/.hydra/config.yaml b/outputs/2023-12-11/10-35-57/.hydra/config.yaml
new file mode 100644
index 0000000..0967ef4
--- /dev/null
+++ b/outputs/2023-12-11/10-35-57/.hydra/config.yaml
@@ -0,0 +1 @@
+{}
diff --git a/outputs/2023-12-11/10-35-57/.hydra/hydra.yaml b/outputs/2023-12-11/10-35-57/.hydra/hydra.yaml
new file mode 100644
index 0000000..3526e0f
--- /dev/null
+++ b/outputs/2023-12-11/10-35-57/.hydra/hydra.yaml
@@ -0,0 +1,154 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task: []
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: ''
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/funasr/cli
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-35-57
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/10-35-57/.hydra/overrides.yaml b/outputs/2023-12-11/10-35-57/.hydra/overrides.yaml
new file mode 100644
index 0000000..fe51488
--- /dev/null
+++ b/outputs/2023-12-11/10-35-57/.hydra/overrides.yaml
@@ -0,0 +1 @@
+[]
diff --git a/outputs/2023-12-11/10-35-57/train_cli.log b/outputs/2023-12-11/10-35-57/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/10-35-57/train_cli.log
diff --git a/outputs/2023-12-11/10-37-01/.hydra/config.yaml b/outputs/2023-12-11/10-37-01/.hydra/config.yaml
new file mode 100644
index 0000000..0967ef4
--- /dev/null
+++ b/outputs/2023-12-11/10-37-01/.hydra/config.yaml
@@ -0,0 +1 @@
+{}
diff --git a/outputs/2023-12-11/10-37-01/.hydra/hydra.yaml b/outputs/2023-12-11/10-37-01/.hydra/hydra.yaml
new file mode 100644
index 0000000..1f352f2
--- /dev/null
+++ b/outputs/2023-12-11/10-37-01/.hydra/hydra.yaml
@@ -0,0 +1,154 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task: []
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: ''
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/funasr/cli
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-37-01
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/10-37-01/.hydra/overrides.yaml b/outputs/2023-12-11/10-37-01/.hydra/overrides.yaml
new file mode 100644
index 0000000..fe51488
--- /dev/null
+++ b/outputs/2023-12-11/10-37-01/.hydra/overrides.yaml
@@ -0,0 +1 @@
+[]
diff --git a/outputs/2023-12-11/10-37-01/train_cli.log b/outputs/2023-12-11/10-37-01/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/10-37-01/train_cli.log
diff --git a/outputs/2023-12-11/10-39-53/.hydra/config.yaml b/outputs/2023-12-11/10-39-53/.hydra/config.yaml
new file mode 100644
index 0000000..0967ef4
--- /dev/null
+++ b/outputs/2023-12-11/10-39-53/.hydra/config.yaml
@@ -0,0 +1 @@
+{}
diff --git a/outputs/2023-12-11/10-39-53/.hydra/hydra.yaml b/outputs/2023-12-11/10-39-53/.hydra/hydra.yaml
new file mode 100644
index 0000000..2be0842
--- /dev/null
+++ b/outputs/2023-12-11/10-39-53/.hydra/hydra.yaml
@@ -0,0 +1,154 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task: []
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: ''
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/funasr/cli
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-39-53
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/10-39-53/.hydra/overrides.yaml b/outputs/2023-12-11/10-39-53/.hydra/overrides.yaml
new file mode 100644
index 0000000..fe51488
--- /dev/null
+++ b/outputs/2023-12-11/10-39-53/.hydra/overrides.yaml
@@ -0,0 +1 @@
+[]
diff --git a/outputs/2023-12-11/10-39-53/train_cli.log b/outputs/2023-12-11/10-39-53/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/10-39-53/train_cli.log
diff --git a/outputs/2023-12-11/10-41-29/.hydra/config.yaml b/outputs/2023-12-11/10-41-29/.hydra/config.yaml
new file mode 100644
index 0000000..5af8fee
--- /dev/null
+++ b/outputs/2023-12-11/10-41-29/.hydra/config.yaml
@@ -0,0 +1,74 @@
+model: funasr.cli.models.paraformer:Paraformer
+model_conf:
+ ctc_weight: 0.3
+ lsm_weight: 0.1
+ length_normalized_loss: false
+ predictor_weight: 1.0
+ sampling_ratio: 0.4
+ use_1st_decoder_loss: true
+frontend: wav_frontend
+frontend_conf:
+ fs: 16000
+ window: hamming
+ n_mels: 80
+ frame_length: 25
+ frame_shift: 10
+ lfr_m: 1
+ lfr_n: 1
+train_conf:
+ accum_grad: 1
+ grad_clip: 5
+ max_epoch: 150
+ val_scheduler_criterion:
+ - valid
+ - acc
+ best_model_criterion:
+ - - valid
+ - acc
+ - max
+ keep_nbest_models: 10
+ log_interval: 50
+optim: adam
+optim_conf:
+ lr: 0.0005
+scheduler: warmuplr
+scheduler_conf:
+ warmup_steps: 30000
+specaug: specaug
+specaug_conf:
+ apply_time_warp: true
+ time_warp_window: 5
+ time_warp_mode: bicubic
+ apply_freq_mask: true
+ freq_mask_width_range:
+ - 0
+ - 30
+ num_freq_mask: 2
+ apply_time_mask: true
+ time_mask_width_range:
+ - 0
+ - 40
+ num_time_mask: 2
+predictor: cif_predictor
+predictor_conf:
+ idim: 256
+ threshold: 1.0
+ l_order: 1
+ r_order: 1
+ tail_threshold: 0.45
+dataset_conf:
+ data_names: speech,text
+ data_types: sound,text
+ shuffle: true
+ shuffle_conf:
+ shuffle_size: 2048
+ sort_size: 500
+ batch_conf:
+ batch_type: example
+ batch_size: 2
+ num_workers: 8
+normalize: null
+model_pretrain: damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/10-41-29/.hydra/hydra.yaml b/outputs/2023-12-11/10-41-29/.hydra/hydra.yaml
new file mode 100644
index 0000000..66deec6
--- /dev/null
+++ b/outputs/2023-12-11/10-41-29/.hydra/hydra.yaml
@@ -0,0 +1,158 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: finetune.yaml
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/examples/industrial_data_pretraining/paraformer-large/conf
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-41-29
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/10-41-29/.hydra/overrides.yaml b/outputs/2023-12-11/10-41-29/.hydra/overrides.yaml
new file mode 100644
index 0000000..d813845
--- /dev/null
+++ b/outputs/2023-12-11/10-41-29/.hydra/overrides.yaml
@@ -0,0 +1,4 @@
+- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/10-41-29/train_cli.log b/outputs/2023-12-11/10-41-29/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/10-41-29/train_cli.log
diff --git a/outputs/2023-12-11/10-44-08/.hydra/config.yaml b/outputs/2023-12-11/10-44-08/.hydra/config.yaml
new file mode 100644
index 0000000..380db42
--- /dev/null
+++ b/outputs/2023-12-11/10-44-08/.hydra/config.yaml
@@ -0,0 +1,4 @@
+model_pretrain: damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/10-44-08/.hydra/hydra.yaml b/outputs/2023-12-11/10-44-08/.hydra/hydra.yaml
new file mode 100644
index 0000000..fd88655
--- /dev/null
+++ b/outputs/2023-12-11/10-44-08/.hydra/hydra.yaml
@@ -0,0 +1,158 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/funasr/cli
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-44-08
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/10-44-08/.hydra/overrides.yaml b/outputs/2023-12-11/10-44-08/.hydra/overrides.yaml
new file mode 100644
index 0000000..d813845
--- /dev/null
+++ b/outputs/2023-12-11/10-44-08/.hydra/overrides.yaml
@@ -0,0 +1,4 @@
+- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/10-44-08/train_cli.log b/outputs/2023-12-11/10-44-08/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/10-44-08/train_cli.log
diff --git a/outputs/2023-12-11/10-45-48/.hydra/config.yaml b/outputs/2023-12-11/10-45-48/.hydra/config.yaml
new file mode 100644
index 0000000..380db42
--- /dev/null
+++ b/outputs/2023-12-11/10-45-48/.hydra/config.yaml
@@ -0,0 +1,4 @@
+model_pretrain: damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/10-45-48/.hydra/hydra.yaml b/outputs/2023-12-11/10-45-48/.hydra/hydra.yaml
new file mode 100644
index 0000000..8d4d4f5
--- /dev/null
+++ b/outputs/2023-12-11/10-45-48/.hydra/hydra.yaml
@@ -0,0 +1,158 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/funasr/cli
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/10-45-48
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/10-45-48/.hydra/overrides.yaml b/outputs/2023-12-11/10-45-48/.hydra/overrides.yaml
new file mode 100644
index 0000000..d813845
--- /dev/null
+++ b/outputs/2023-12-11/10-45-48/.hydra/overrides.yaml
@@ -0,0 +1,4 @@
+- +model_pretrain=damo/speech_paraformer-large-vad-punc_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/10-45-48/train_cli.log b/outputs/2023-12-11/10-45-48/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/10-45-48/train_cli.log
diff --git a/outputs/2023-12-11/11-26-16/.hydra/config.yaml b/outputs/2023-12-11/11-26-16/.hydra/config.yaml
new file mode 100644
index 0000000..10ea676
--- /dev/null
+++ b/outputs/2023-12-11/11-26-16/.hydra/config.yaml
@@ -0,0 +1,4 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/11-26-16/.hydra/hydra.yaml b/outputs/2023-12-11/11-26-16/.hydra/hydra.yaml
new file mode 100644
index 0000000..e3c206b
--- /dev/null
+++ b/outputs/2023-12-11/11-26-16/.hydra/hydra.yaml
@@ -0,0 +1,158 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.1'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: /Users/zhifu/funasr_github/funasr/cli
+ schema: file
+ provider: main
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-26-16
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-26-16/.hydra/overrides.yaml b/outputs/2023-12-11/11-26-16/.hydra/overrides.yaml
new file mode 100644
index 0000000..95db593
--- /dev/null
+++ b/outputs/2023-12-11/11-26-16/.hydra/overrides.yaml
@@ -0,0 +1,4 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/11-26-16/train_cli.log b/outputs/2023-12-11/11-26-16/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/11-26-16/train_cli.log
diff --git a/outputs/2023-12-11/11-32-18/.hydra/config.yaml b/outputs/2023-12-11/11-32-18/.hydra/config.yaml
new file mode 100644
index 0000000..10ea676
--- /dev/null
+++ b/outputs/2023-12-11/11-32-18/.hydra/config.yaml
@@ -0,0 +1,4 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/11-32-18/.hydra/hydra.yaml b/outputs/2023-12-11/11-32-18/.hydra/hydra.yaml
new file mode 100644
index 0000000..b9ddf5b
--- /dev/null
+++ b/outputs/2023-12-11/11-32-18/.hydra/hydra.yaml
@@ -0,0 +1,155 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-32-18
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-32-18/.hydra/overrides.yaml b/outputs/2023-12-11/11-32-18/.hydra/overrides.yaml
new file mode 100644
index 0000000..95db593
--- /dev/null
+++ b/outputs/2023-12-11/11-32-18/.hydra/overrides.yaml
@@ -0,0 +1,4 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
diff --git a/outputs/2023-12-11/11-32-18/train_cli.log b/outputs/2023-12-11/11-32-18/train_cli.log
new file mode 100644
index 0000000..9dca346
--- /dev/null
+++ b/outputs/2023-12-11/11-32-18/train_cli.log
@@ -0,0 +1 @@
+[2023-12-11 11:32:54,061][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
diff --git a/outputs/2023-12-11/11-33-27/.hydra/config.yaml b/outputs/2023-12-11/11-33-27/.hydra/config.yaml
new file mode 100644
index 0000000..2cd5632
--- /dev/null
+++ b/outputs/2023-12-11/11-33-27/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/11-33-27/.hydra/hydra.yaml b/outputs/2023-12-11/11-33-27/.hydra/hydra.yaml
new file mode 100644
index 0000000..67e44bf
--- /dev/null
+++ b/outputs/2023-12-11/11-33-27/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-33-27
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-33-27/.hydra/overrides.yaml b/outputs/2023-12-11/11-33-27/.hydra/overrides.yaml
new file mode 100644
index 0000000..1337501
--- /dev/null
+++ b/outputs/2023-12-11/11-33-27/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/11-33-27/train_cli.log b/outputs/2023-12-11/11-33-27/train_cli.log
new file mode 100644
index 0000000..9fabe15
--- /dev/null
+++ b/outputs/2023-12-11/11-33-27/train_cli.log
@@ -0,0 +1 @@
+[2023-12-11 11:33:45,313][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
diff --git a/outputs/2023-12-11/11-36-25/.hydra/config.yaml b/outputs/2023-12-11/11-36-25/.hydra/config.yaml
new file mode 100644
index 0000000..2cd5632
--- /dev/null
+++ b/outputs/2023-12-11/11-36-25/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /nfs/zhifu.gzf/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/11-36-25/.hydra/hydra.yaml b/outputs/2023-12-11/11-36-25/.hydra/hydra.yaml
new file mode 100644
index 0000000..c4cf67f
--- /dev/null
+++ b/outputs/2023-12-11/11-36-25/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-36-25
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-36-25/.hydra/overrides.yaml b/outputs/2023-12-11/11-36-25/.hydra/overrides.yaml
new file mode 100644
index 0000000..1337501
--- /dev/null
+++ b/outputs/2023-12-11/11-36-25/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/nfs/zhifu.gzf/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/11-36-25/train_cli.log b/outputs/2023-12-11/11-36-25/train_cli.log
new file mode 100644
index 0000000..cf0585d
--- /dev/null
+++ b/outputs/2023-12-11/11-36-25/train_cli.log
@@ -0,0 +1,2 @@
+[2023-12-11 11:36:30,624][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 11:36:31,819][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/11-37-13/.hydra/config.yaml b/outputs/2023-12-11/11-37-13/.hydra/config.yaml
new file mode 100644
index 0000000..2f51012
--- /dev/null
+++ b/outputs/2023-12-11/11-37-13/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu.gzf/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/11-37-13/.hydra/hydra.yaml b/outputs/2023-12-11/11-37-13/.hydra/hydra.yaml
new file mode 100644
index 0000000..aaaf4c2
--- /dev/null
+++ b/outputs/2023-12-11/11-37-13/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu.gzf/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu.gzf/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-37-13
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-37-13/.hydra/overrides.yaml b/outputs/2023-12-11/11-37-13/.hydra/overrides.yaml
new file mode 100644
index 0000000..7a67075
--- /dev/null
+++ b/outputs/2023-12-11/11-37-13/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu.gzf/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/11-37-13/train_cli.log b/outputs/2023-12-11/11-37-13/train_cli.log
new file mode 100644
index 0000000..0d78f1b
--- /dev/null
+++ b/outputs/2023-12-11/11-37-13/train_cli.log
@@ -0,0 +1,2 @@
+[2023-12-11 11:43:26,900][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 11:43:28,209][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/11-44-36/.hydra/config.yaml b/outputs/2023-12-11/11-44-36/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/11-44-36/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/11-44-36/.hydra/hydra.yaml b/outputs/2023-12-11/11-44-36/.hydra/hydra.yaml
new file mode 100644
index 0000000..225a367
--- /dev/null
+++ b/outputs/2023-12-11/11-44-36/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-44-36
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-44-36/.hydra/overrides.yaml b/outputs/2023-12-11/11-44-36/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/11-44-36/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/11-44-36/train_cli.log b/outputs/2023-12-11/11-44-36/train_cli.log
new file mode 100644
index 0000000..5ef50d6
--- /dev/null
+++ b/outputs/2023-12-11/11-44-36/train_cli.log
@@ -0,0 +1,2 @@
+[2023-12-11 11:44:48,432][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 11:44:49,667][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/11-48-01/.hydra/config.yaml b/outputs/2023-12-11/11-48-01/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/11-48-01/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/11-48-01/.hydra/hydra.yaml b/outputs/2023-12-11/11-48-01/.hydra/hydra.yaml
new file mode 100644
index 0000000..3579e9e
--- /dev/null
+++ b/outputs/2023-12-11/11-48-01/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/11-48-01
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/11-48-01/.hydra/overrides.yaml b/outputs/2023-12-11/11-48-01/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/11-48-01/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/11-48-01/train_cli.log b/outputs/2023-12-11/11-48-01/train_cli.log
new file mode 100644
index 0000000..3531939
--- /dev/null
+++ b/outputs/2023-12-11/11-48-01/train_cli.log
@@ -0,0 +1,3 @@
+[2023-12-11 11:48:05,427][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 11:48:07,007][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 11:48:07,026][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-06-39/.hydra/config.yaml b/outputs/2023-12-11/13-06-39/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-06-39/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-06-39/.hydra/hydra.yaml b/outputs/2023-12-11/13-06-39/.hydra/hydra.yaml
new file mode 100644
index 0000000..e92f2f8
--- /dev/null
+++ b/outputs/2023-12-11/13-06-39/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-06-39
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-06-39/.hydra/overrides.yaml b/outputs/2023-12-11/13-06-39/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-06-39/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-06-39/train_cli.log b/outputs/2023-12-11/13-06-39/train_cli.log
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/outputs/2023-12-11/13-06-39/train_cli.log
diff --git a/outputs/2023-12-11/13-07-13/.hydra/config.yaml b/outputs/2023-12-11/13-07-13/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-07-13/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-07-13/.hydra/hydra.yaml b/outputs/2023-12-11/13-07-13/.hydra/hydra.yaml
new file mode 100644
index 0000000..f47436a
--- /dev/null
+++ b/outputs/2023-12-11/13-07-13/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-07-13
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-07-13/.hydra/overrides.yaml b/outputs/2023-12-11/13-07-13/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-07-13/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-07-13/train_cli.log b/outputs/2023-12-11/13-07-13/train_cli.log
new file mode 100644
index 0000000..74bb187
--- /dev/null
+++ b/outputs/2023-12-11/13-07-13/train_cli.log
@@ -0,0 +1,3 @@
+[2023-12-11 13:07:30,599][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:07:31,757][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:07:31,859][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-12-46/.hydra/config.yaml b/outputs/2023-12-11/13-12-46/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-12-46/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-12-46/.hydra/hydra.yaml b/outputs/2023-12-11/13-12-46/.hydra/hydra.yaml
new file mode 100644
index 0000000..c0a361e
--- /dev/null
+++ b/outputs/2023-12-11/13-12-46/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-12-46
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-12-46/.hydra/overrides.yaml b/outputs/2023-12-11/13-12-46/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-12-46/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-12-46/train_cli.log b/outputs/2023-12-11/13-12-46/train_cli.log
new file mode 100644
index 0000000..1aed15c
--- /dev/null
+++ b/outputs/2023-12-11/13-12-46/train_cli.log
@@ -0,0 +1,5 @@
+[2023-12-11 13:12:51,157][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
+[2023-12-11 13:12:51,157][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:12:52,231][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:12:52,266][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
+[2023-12-11 13:12:52,266][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-16-01/.hydra/config.yaml b/outputs/2023-12-11/13-16-01/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-16-01/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-16-01/.hydra/hydra.yaml b/outputs/2023-12-11/13-16-01/.hydra/hydra.yaml
new file mode 100644
index 0000000..dc32784
--- /dev/null
+++ b/outputs/2023-12-11/13-16-01/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-16-01
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-16-01/.hydra/overrides.yaml b/outputs/2023-12-11/13-16-01/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-16-01/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-16-01/train_cli.log b/outputs/2023-12-11/13-16-01/train_cli.log
new file mode 100644
index 0000000..66787f9
--- /dev/null
+++ b/outputs/2023-12-11/13-16-01/train_cli.log
@@ -0,0 +1,6 @@
+[2023-12-11 13:16:27,835][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
+[2023-12-11 13:16:27,835][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:16:28,828][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:16:28,828][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
+[2023-12-11 13:16:28,835][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
+[2023-12-11 13:16:28,835][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-19-05/.hydra/config.yaml b/outputs/2023-12-11/13-19-05/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-19-05/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-19-05/.hydra/hydra.yaml b/outputs/2023-12-11/13-19-05/.hydra/hydra.yaml
new file mode 100644
index 0000000..d89b6c1
--- /dev/null
+++ b/outputs/2023-12-11/13-19-05/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-19-05
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-19-05/.hydra/overrides.yaml b/outputs/2023-12-11/13-19-05/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-19-05/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-19-05/train_cli.log b/outputs/2023-12-11/13-19-05/train_cli.log
new file mode 100644
index 0000000..1832e8e
--- /dev/null
+++ b/outputs/2023-12-11/13-19-05/train_cli.log
@@ -0,0 +1,6 @@
+[2023-12-11 13:19:09,868][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
+[2023-12-11 13:19:09,868][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:19:11,006][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:19:11,006][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
+[2023-12-11 13:19:11,014][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
+[2023-12-11 13:19:11,014][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-24-15/.hydra/config.yaml b/outputs/2023-12-11/13-24-15/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-24-15/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-24-15/.hydra/hydra.yaml b/outputs/2023-12-11/13-24-15/.hydra/hydra.yaml
new file mode 100644
index 0000000..37decbb
--- /dev/null
+++ b/outputs/2023-12-11/13-24-15/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-24-15
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-24-15/.hydra/overrides.yaml b/outputs/2023-12-11/13-24-15/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-24-15/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-24-15/train_cli.log b/outputs/2023-12-11/13-24-15/train_cli.log
new file mode 100644
index 0000000..06122f2
--- /dev/null
+++ b/outputs/2023-12-11/13-24-15/train_cli.log
@@ -0,0 +1,6 @@
+[2023-12-11 13:24:21,397][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
+[2023-12-11 13:24:21,397][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:24:21,980][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:24:21,980][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
+[2023-12-11 13:24:21,987][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
+[2023-12-11 13:24:21,987][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-37-51/.hydra/config.yaml b/outputs/2023-12-11/13-37-51/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-37-51/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-37-51/.hydra/hydra.yaml b/outputs/2023-12-11/13-37-51/.hydra/hydra.yaml
new file mode 100644
index 0000000..7a168d7
--- /dev/null
+++ b/outputs/2023-12-11/13-37-51/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-37-51
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-37-51/.hydra/overrides.yaml b/outputs/2023-12-11/13-37-51/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-37-51/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-37-51/train_cli.log b/outputs/2023-12-11/13-37-51/train_cli.log
new file mode 100644
index 0000000..724cc0d
--- /dev/null
+++ b/outputs/2023-12-11/13-37-51/train_cli.log
@@ -0,0 +1,6 @@
+[2023-12-11 13:38:00,018][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
+[2023-12-11 13:38:00,019][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:38:01,162][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:38:01,162][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
+[2023-12-11 13:38:01,248][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
+[2023-12-11 13:38:01,249][root][WARNING] - distributed is not initialized, only single shard
diff --git a/outputs/2023-12-11/13-39-37/.hydra/config.yaml b/outputs/2023-12-11/13-39-37/.hydra/config.yaml
new file mode 100644
index 0000000..85ff69f
--- /dev/null
+++ b/outputs/2023-12-11/13-39-37/.hydra/config.yaml
@@ -0,0 +1,5 @@
+model_pretrain: /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+token_list: /Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+train_data_set_list: /Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+output_dir: /Users/zhifu/Downloads/ckpt/funasr2/exp2
+device: cpu
diff --git a/outputs/2023-12-11/13-39-37/.hydra/hydra.yaml b/outputs/2023-12-11/13-39-37/.hydra/hydra.yaml
new file mode 100644
index 0000000..85aae43
--- /dev/null
+++ b/outputs/2023-12-11/13-39-37/.hydra/hydra.yaml
@@ -0,0 +1,156 @@
+hydra:
+ run:
+ dir: outputs/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ sweep:
+ dir: multirun/${now:%Y-%m-%d}/${now:%H-%M-%S}
+ subdir: ${hydra.job.num}
+ launcher:
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
+ sweeper:
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
+ max_batch_size: null
+ params: null
+ help:
+ app_name: ${hydra.job.name}
+ header: '${hydra.help.app_name} is powered by Hydra.
+
+ '
+ footer: 'Powered by Hydra (https://hydra.cc)
+
+ Use --hydra-help to view Hydra specific help
+
+ '
+ template: '${hydra.help.header}
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (group=option)
+
+
+ $APP_CONFIG_GROUPS
+
+
+ == Config ==
+
+ Override anything in the config (foo.bar=value)
+
+
+ $CONFIG
+
+
+ ${hydra.help.footer}
+
+ '
+ hydra_help:
+ template: 'Hydra (${hydra.runtime.version})
+
+ See https://hydra.cc for more info.
+
+
+ == Flags ==
+
+ $FLAGS_HELP
+
+
+ == Configuration groups ==
+
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
+ to command line)
+
+
+ $HYDRA_CONFIG_GROUPS
+
+
+ Use ''--cfg hydra'' to Show the Hydra config.
+
+ '
+ hydra_help: ???
+ hydra_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][HYDRA] %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ root:
+ level: INFO
+ handlers:
+ - console
+ loggers:
+ logging_example:
+ level: DEBUG
+ disable_existing_loggers: false
+ job_logging:
+ version: 1
+ formatters:
+ simple:
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
+ handlers:
+ console:
+ class: logging.StreamHandler
+ formatter: simple
+ stream: ext://sys.stdout
+ file:
+ class: logging.FileHandler
+ formatter: simple
+ filename: ${hydra.runtime.output_dir}/${hydra.job.name}.log
+ root:
+ level: INFO
+ handlers:
+ - console
+ - file
+ disable_existing_loggers: false
+ env: {}
+ mode: RUN
+ searchpath: []
+ callbacks: {}
+ output_subdir: .hydra
+ overrides:
+ hydra:
+ - hydra.mode=RUN
+ task:
+ - +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+ - +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+ - +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ - +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+ - +device=cpu
+ job:
+ name: train_cli
+ chdir: null
+ override_dirname: +device=cpu,+model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch,+output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2,+token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt,+train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+ id: ???
+ num: ???
+ config_name: null
+ env_set: {}
+ env_copy: []
+ config:
+ override_dirname:
+ kv_sep: '='
+ item_sep: ','
+ exclude_keys: []
+ runtime:
+ version: 1.3.2
+ version_base: '1.3'
+ cwd: /Users/zhifu/funasr_github
+ config_sources:
+ - path: hydra.conf
+ schema: pkg
+ provider: hydra
+ - path: ''
+ schema: structured
+ provider: schema
+ output_dir: /Users/zhifu/funasr_github/outputs/2023-12-11/13-39-37
+ choices:
+ hydra/env: default
+ hydra/callbacks: null
+ hydra/job_logging: default
+ hydra/hydra_logging: default
+ hydra/hydra_help: default
+ hydra/help: default
+ hydra/sweeper: basic
+ hydra/launcher: basic
+ hydra/output: default
+ verbose: false
diff --git a/outputs/2023-12-11/13-39-37/.hydra/overrides.yaml b/outputs/2023-12-11/13-39-37/.hydra/overrides.yaml
new file mode 100644
index 0000000..983d97f
--- /dev/null
+++ b/outputs/2023-12-11/13-39-37/.hydra/overrides.yaml
@@ -0,0 +1,5 @@
+- +model_pretrain=/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch
+- +token_list=/Users/zhifu/.cache/modelscope/hub/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/tokens.txt
+- +train_data_set_list=/Users/zhifu/funasr_github/test_local/aishell2_dev_ios/asr_task_debug_len.jsonl
+- +output_dir=/Users/zhifu/Downloads/ckpt/funasr2/exp2
+- +device=cpu
diff --git a/outputs/2023-12-11/13-39-37/train_cli.log b/outputs/2023-12-11/13-39-37/train_cli.log
new file mode 100644
index 0000000..4238628
--- /dev/null
+++ b/outputs/2023-12-11/13-39-37/train_cli.log
@@ -0,0 +1,6 @@
+[2023-12-11 13:39:40,353][root][INFO] - init_param is not None: ('/Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb',)
+[2023-12-11 13:39:40,353][root][INFO] - Loading pretrained params from /Users/zhifu/modelscope_models/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/model.pb
+[2023-12-11 13:39:41,780][root][WARNING] - distributed is not initialized, only single shard
+[2023-12-11 13:39:41,780][root][INFO] - in rank: 0, num of samplers: 640, total_num of samplers across ranks: 640
+[2023-12-11 13:39:41,861][root][INFO] - config.yaml is saved to: /Users/zhifu/Downloads/ckpt/funasr2/exp2/config.yaml
+[2023-12-11 13:39:41,862][root][WARNING] - distributed is not initialized, only single shard
--
Gitblit v1.9.1