zhifu gao
2023-02-20 0856ea2ebdcb976db6e786de5cd79fae3d35cd4c
funasr/train/distributed_utils.py
@@ -60,8 +60,6 @@
            self.dist_rank = torch.distributed.get_rank()
            self.dist_world_size = torch.distributed.get_world_size()
            self.local_rank = args.local_rank
            logging.info("world size: {}, rank: {}, local_rank: {}".format(self.dist_world_size, self.dist_rank,
                                                                           self.local_rank))
    def init_options_pai(self):
        if self.distributed:
@@ -119,8 +117,6 @@
            self.dist_rank = torch.distributed.get_rank()
            self.dist_world_size = torch.distributed.get_world_size()
            self.local_rank = args.local_rank
            logging.info("world size: {}, rank: {}, local_rank: {}".format(self.dist_world_size, self.dist_rank,
                                                                           self.local_rank))
def resolve_distributed_mode(args):