Merge pull request #2267 from modelscope/dev_sx2
fix model_wrapper and seaco finetune
| | |
| | | ++train_conf.avg_nbest_model=10 \ |
| | | ++train_conf.use_deepspeed=false \ |
| | | ++train_conf.deepspeed_config=${deepspeed_config} \ |
| | | ++train_conf.find_unused_parameters=true \ |
| | | ++optim_conf.lr=0.0002 \ |
| | | ++output_dir="${output_dir}" &> ${log_file} |
| | |
| | | **kwargs.get("train_conf"), |
| | | ) |
| | | |
| | | model = trainer.warp_model(model) |
| | | model = trainer.warp_model(model, **kwargs) |
| | | |
| | | kwargs["device"] = int(os.environ.get("LOCAL_RANK", 0)) |
| | | trainer.device = int(os.environ.get("LOCAL_RANK", 0)) |