From de6da13e1901cc6def6f3ca2b689621e936eb04a Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Mon, 22 May 2023 12:12:01 -0400 Subject: [PATCH] don't need to set here --- src/axolotl/utils/trainer.py | 3 --- 1 file changed, 3 deletions(-) diff --git a/src/axolotl/utils/trainer.py b/src/axolotl/utils/trainer.py index 4c6eb7626..4336f740c 100644 --- a/src/axolotl/utils/trainer.py +++ b/src/axolotl/utils/trainer.py @@ -81,9 +81,6 @@ def setup_trainer(cfg, train_dataset, eval_dataset, model, tokenizer): training_arguments_kwargs["fsdp"] = cfg.fsdp if cfg.fsdp_config: training_arguments_kwargs["fsdp_config"] = dict(cfg.fsdp_config) - # can't set optimizers directly on trainer when using fsdp, so set them here - if cfg.optimizer: - training_arguments_kwargs["optim"] = cfg.optimizer # deepspeed if (