use wip branch for expected 4.48.1
This commit is contained in:
@@ -14,7 +14,7 @@ packaging==23.2
|
||||
|
||||
peft==0.14.0
|
||||
# transformers==4.48.1
|
||||
transformers @ git+https://github.com/huggingface/transformers.git@v4.48-release
|
||||
transformers @ git+https://github.com/winglian/transformers.git@pre-v4.48.1
|
||||
tokenizers>=0.21.0
|
||||
accelerate==1.2.1
|
||||
datasets==3.2.0
|
||||
|
||||
@@ -386,15 +386,15 @@ class ModelLoader:
|
||||
if self.cfg.flash_attention:
|
||||
self.patch_attention()
|
||||
|
||||
if self.cfg.model_config_type == "llama":
|
||||
from axolotl.monkeypatch.trainer_grad_accum import ( # patch_forward_for_ga,
|
||||
patch_flash_attention_forward,
|
||||
patch_training_step_for_ga,
|
||||
)
|
||||
|
||||
patch_flash_attention_forward()
|
||||
# patch_forward_for_ga()
|
||||
patch_training_step_for_ga()
|
||||
# if self.cfg.model_config_type == "llama":
|
||||
# from axolotl.monkeypatch.trainer_grad_accum import ( # patch_forward_for_ga,
|
||||
# patch_flash_attention_forward,
|
||||
# patch_training_step_for_ga,
|
||||
# )
|
||||
#
|
||||
# patch_flash_attention_forward()
|
||||
# # patch_forward_for_ga()
|
||||
# patch_training_step_for_ga()
|
||||
|
||||
if self.cfg.sample_packing and self.cfg.s2_attention:
|
||||
raise ValueError(
|
||||
|
||||
Reference in New Issue
Block a user