嘉渊
2023-04-24 189b51d42bd29032091f1e29ae5585eb52c0af57
update
3个文件已修改
10 ■■■■ 已修改文件
egs/aishell/paraformer/run.sh 1 ●●●● 补丁 | 查看 | 原始文档 | blame | 历史
funasr/bin/train.py 7 ●●●●● 补丁 | 查看 | 原始文档 | blame | 历史
funasr/datasets/small_datasets/sequence_iter_factory.py 2 ●●● 补丁 | 查看 | 原始文档 | blame | 历史
egs/aishell/paraformer/run.sh
@@ -180,7 +180,6 @@
                --input_size $feats_dim \
                --ngpu $gpu_num \
                --num_worker_count $count \
                --multiprocessing_distributed true \
                --dist_init_method $init_method \
                --dist_world_size $world_size \
                --dist_rank $rank \
funasr/bin/train.py
@@ -77,6 +77,12 @@
        help="Whether to use the find_unused_parameters in "
             "torch.nn.parallel.DistributedDataParallel ",
    )
    parser.add_argument(
        "--gpu_id",
        type=int,
        default=0,
        help="local gpu id.",
    )
    # cudnn related
    parser.add_argument(
@@ -399,6 +405,7 @@
    torch.backends.cudnn.deterministic = args.cudnn_deterministic
    # ddp init
    os.environ['CUDA_VISIBLE_DEVICES'] = str(args.gpu_id)
    args.distributed = args.dist_world_size > 1
    distributed_option = build_distributed(args)
funasr/datasets/small_datasets/sequence_iter_factory.py
@@ -62,7 +62,7 @@
        # sampler
        dataset_conf = args.dataset_conf
        batch_sampler = LengthBatchSampler(
            batch_bins=dataset_conf["batch_size"],
            batch_bins=dataset_conf["batch_size"] * args.ngpu,
            shape_files=shape_files,
            sort_in_batch=dataset_conf["sort_in_batch"] if hasattr(dataset_conf, "sort_in_batch") else "descending",
            sort_batch=dataset_conf["sort_batch"] if hasattr(dataset_conf, "sort_batch") else "ascending",