diff --git a/requirements.txt b/requirements.txt index d561b4326..6dac3c7a2 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,10 +1,10 @@ --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/ packaging==23.2 peft==0.13.2 -transformers==4.46.0 +transformers==4.46.2 tokenizers>=0.20.1 bitsandbytes==0.44.1 -accelerate==1.0.1 +accelerate==1.0.2 datasets==3.0.1 deepspeed==0.15.3 pydantic==2.6.3 @@ -34,7 +34,7 @@ tensorboard python-dotenv==1.0.1 autoawq>=0.2.5 triton>=2.3.0 -liger-kernel==0.3.1 +liger-kernel==0.4.0 mamba-ssm==1.2.0.post1 diff --git a/src/axolotl/core/trainer_builder.py b/src/axolotl/core/trainer_builder.py index aab9a80b8..f4a2f9001 100755 --- a/src/axolotl/core/trainer_builder.py +++ b/src/axolotl/core/trainer_builder.py @@ -896,13 +896,13 @@ class AxolotlTrainer(SchedulerMixin, Trainer): for key, value in metrics.items(): self._stored_metrics[train_eval][key].append(value) - def _save_checkpoint(self, model, trial, metrics=None): + def _save_checkpoint(self, model, trial): # make sure the checkpoint dir exists, since trainer is flakey checkpoint_folder = f"{PREFIX_CHECKPOINT_DIR}-{self.state.global_step}" run_dir = self._get_output_dir(trial=trial) output_dir = os.path.join(run_dir, checkpoint_folder) os.makedirs(output_dir, exist_ok=True) - return super()._save_checkpoint(model, trial, metrics=metrics) + return super()._save_checkpoint(model, trial) class AxolotlMambaTrainer(AxolotlTrainer):