From 54931dd4e1a099d7d6f144c4e12e5453deb3aa26 Mon Sep 17 00:00:00 2001
From: 雾聪 <wucong.lyb@alibaba-inc.com>
Date: 星期三, 28 六月 2023 10:41:57 +0800
Subject: [PATCH] Merge branch 'main' of https://github.com/alibaba-damo-academy/FunASR into main
---
funasr/datasets/small_datasets/sequence_iter_factory.py | 10 +++++-----
1 files changed, 5 insertions(+), 5 deletions(-)
diff --git a/funasr/datasets/small_datasets/sequence_iter_factory.py b/funasr/datasets/small_datasets/sequence_iter_factory.py
index c69b2ba..3ebcc5a 100644
--- a/funasr/datasets/small_datasets/sequence_iter_factory.py
+++ b/funasr/datasets/small_datasets/sequence_iter_factory.py
@@ -27,8 +27,7 @@
class SequenceIterFactory(AbsIterFactory):
- """Build iterator for each epoch.
-
+ """Build iterator for each epoch, modified from ESPnet
"""
@@ -58,12 +57,13 @@
data_path_and_name_and_type,
preprocess=preprocess_fn,
dest_sample_rate=dest_sample_rate,
+ speed_perturb=args.speed_perturb if mode=="train" else None,
)
# sampler
dataset_conf = args.dataset_conf
batch_sampler = LengthBatchSampler(
- batch_bins=dataset_conf["batch_size"],
+ batch_bins=dataset_conf["batch_conf"]["batch_size"] * args.ngpu,
shape_files=shape_files,
sort_in_batch=dataset_conf["sort_in_batch"] if hasattr(dataset_conf, "sort_in_batch") else "descending",
sort_batch=dataset_conf["sort_batch"] if hasattr(dataset_conf, "sort_batch") else "ascending",
@@ -84,7 +84,7 @@
args.max_update = len(bs_list) * args.max_epoch
logging.info("Max update: {}".format(args.max_update))
- if args.distributed:
+ if args.distributed and mode=="train":
world_size = torch.distributed.get_world_size()
rank = torch.distributed.get_rank()
for batch in batches:
@@ -104,7 +104,7 @@
self.num_iters_per_epoch = None
self.shuffle = mode == "train"
self.seed = args.seed
- self.num_workers = args.num_workers
+ self.num_workers = args.dataset_conf.get("num_workers", 8)
self.collate_fn = collate_fn
self.pin_memory = args.ngpu > 0
--
Gitblit v1.9.1