From e2b3edec45fe1bfd76493acc366df971e89e7ae2 Mon Sep 17 00:00:00 2001
From: Xian Shi <40013335+R1ckShi@users.noreply.github.com>
Date: 星期四, 10 八月 2023 17:30:34 +0800
Subject: [PATCH] Merge pull request #830 from alibaba-damo-academy/dev_ts
---
funasr/datasets/small_datasets/sequence_iter_factory.py | 7 ++++---
1 files changed, 4 insertions(+), 3 deletions(-)
diff --git a/funasr/datasets/small_datasets/sequence_iter_factory.py b/funasr/datasets/small_datasets/sequence_iter_factory.py
index f524c78..8a7279a 100644
--- a/funasr/datasets/small_datasets/sequence_iter_factory.py
+++ b/funasr/datasets/small_datasets/sequence_iter_factory.py
@@ -57,7 +57,7 @@
data_path_and_name_and_type,
preprocess=preprocess_fn,
dest_sample_rate=dest_sample_rate,
- speed_perturb=args.speed_perturb,
+ speed_perturb=args.speed_perturb if mode == "train" else None,
)
# sampler
@@ -66,8 +66,9 @@
batch_bins=dataset_conf["batch_conf"]["batch_size"] * args.ngpu,
shape_files=shape_files,
sort_in_batch=dataset_conf["sort_in_batch"] if hasattr(dataset_conf, "sort_in_batch") else "descending",
- sort_batch=dataset_conf["sort_batch"] if hasattr(dataset_conf, "sort_batch") else "ascending",
+ sort_batch=dataset_conf["sort_batch"] if hasattr(dataset_conf, "sort_batch") else "descending",
drop_last=False,
+ min_batch_size=torch.distributed.get_world_size(),
padding=True,
)
@@ -84,7 +85,7 @@
args.max_update = len(bs_list) * args.max_epoch
logging.info("Max update: {}".format(args.max_update))
- if args.distributed and mode=="train":
+ if args.distributed and mode == "train":
world_size = torch.distributed.get_world_size()
rank = torch.distributed.get_rank()
for batch in batches:
--
Gitblit v1.9.1