diff --git a/requirements.txt b/requirements.txt index 46d0691b6..8f9f55262 100644 --- a/requirements.txt +++ b/requirements.txt @@ -4,7 +4,7 @@ peft==0.13.2 transformers==4.45.2 tokenizers>=0.20.1 bitsandbytes==0.44.1 -accelerate==0.34.2 +accelerate==1.0.1 datasets==3.0.1 deepspeed==0.14.4 pydantic==2.6.3 diff --git a/src/axolotl/train.py b/src/axolotl/train.py index 6ad373655..4ce28d8a3 100644 --- a/src/axolotl/train.py +++ b/src/axolotl/train.py @@ -10,7 +10,6 @@ from typing import Optional, Tuple, Union import torch import transformers.modelcard -from accelerate import Accelerator from accelerate.logging import get_logger from accelerate.utils import save_fsdp_model from datasets import Dataset @@ -97,8 +96,6 @@ def train( if cfg.adapter: msg += " and peft_config..." LOG.debug(msg) - # we wait unitl the last possible moment to setup Accelerator - Accelerator() model, peft_config = load_model( cfg, tokenizer, processor=processor, inference=cli_args.inference )