Compare commits

..

2 Commits

Author SHA1 Message Date
Wing Lian
79612da5c8 perform flakey patched tests in individual runner 2024-12-12 23:22:28 -05:00
Wing Lian
effc4dc409 pin to 4.47.0 (#2180) 2024-12-12 20:17:12 -05:00
11 changed files with 6 additions and 192 deletions

View File

@@ -8,3 +8,8 @@ pytest -v --durations=10 -n8 --ignore=tests/e2e/ --ignore=tests/patched/ /worksp
pytest -v --durations=10 -n1 --dist loadfile /workspace/axolotl/tests/e2e/patched/
pytest -v --durations=10 -n1 --dist loadfile /workspace/axolotl/tests/e2e/integrations/
pytest -v --durations=10 --ignore=tests/e2e/patched/ --ignore=tests/e2e/multigpu/ --ignore=tests/e2e/integrations/ /workspace/axolotl/tests/e2e/
tests=$(pytest --collect-only -q tests/e2e/each)
for t in $tests; do
pytest $t
done

View File

@@ -12,7 +12,7 @@ liger-kernel==0.4.2
packaging==23.2
peft==0.14.0
transformers>=4.46.3
transformers==4.47.0
tokenizers>=0.20.1
accelerate==1.2.0
datasets==3.1.0

View File

@@ -996,15 +996,6 @@ class AxolotlTrainer(SchedulerMixin, Trainer):
os.makedirs(output_dir, exist_ok=True)
return super()._save_checkpoint(model, trial, **kwargs)
def _evaluate(self, *args, **kwargs):
metrics = super()._evaluate(*args, **kwargs)
# cleanup memory after evals
gc.collect()
torch.cuda.empty_cache()
return metrics
class AxolotlMambaTrainer(AxolotlTrainer):
"""

View File

@@ -1,170 +0,0 @@
import contextlib
import inspect
import types
from torchtune.training import OffloadActivations
from transformers import LlamaConfig, LlamaForCausalLM
from axolotl.monkeypatch.unsloth_ import detab_code
HF_MODEL_OUTPUTS = """
outputs = self.model(
input_ids=input_ids,
attention_mask=attention_mask,
position_ids=position_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
use_cache=use_cache,
output_attentions=output_attentions,
output_hidden_states=output_hidden_states,
return_dict=return_dict,
cache_position=cache_position,
**kwargs,
)
""".lstrip()
PATCHED_HF_MODEL_OUTPUTS = """
with self.act_offloading_ctx_manager:
outputs = self.model(
input_ids=input_ids,
attention_mask=attention_mask,
position_ids=position_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
use_cache=use_cache,
output_attentions=output_attentions,
output_hidden_states=output_hidden_states,
return_dict=return_dict,
cache_position=cache_position,
**kwargs,
)
""".lstrip()
LCE_MODEL_OUTPUTS = """
outputs = self.model(
input_ids=input_ids,
attention_mask=attention_mask,
position_ids=position_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
use_cache=use_cache,
output_attentions=output_attentions,
output_hidden_states=output_hidden_states,
return_dict=return_dict,
cache_position=cache_position,
)
""".lstrip()
PATCHED_LCE_OUTPUTS = """
with self.act_offloading_ctx_manager:
outputs = self.model(
input_ids=input_ids,
attention_mask=attention_mask,
position_ids=position_ids,
past_key_values=past_key_values,
inputs_embeds=inputs_embeds,
use_cache=use_cache,
output_attentions=output_attentions,
output_hidden_states=output_hidden_states,
return_dict=return_dict,
cache_position=cache_position,
)
""".lstrip()
HF_GA_FORWARD_1 = """
return_dict = return_dict if return_dict is not None else self.config.use_return_dict
# decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn)
""".lstrip()
PATCHED_HF_GA_FORWARD_1 = """
return_dict = return_dict if return_dict is not None else self.config.use_return_dict
# remove num_items_in_batch otherwise self.model attempts to pass it to flash_attention
num_items_in_batch = kwargs.pop("num_items_in_batch", None)
# decoder outputs consists of (dec_features, layer_state, dec_hidden, dec_attn)
""".lstrip()
HF_GA_FORWARD_2 = """
loss = None
if labels is not None:
loss = self.loss_function(logits=logits, labels=labels, vocab_size=self.config.vocab_size, **kwargs)
""".lstrip()
PATCHED_HF_GA_FORWARD_2 = """
loss = None
if labels is not None:
loss = self.loss_function(logits=logits, labels=labels, vocab_size=self.config.vocab_size, num_items_in_batch=num_items_in_batch, **kwargs)
""".lstrip()
class AxolotlLlamaForCausalLM(LlamaForCausalLM):
act_offloading_ctx_manager = contextlib.nullcontext()
def __init__(self, config: LlamaConfig):
super().__init__(config)
@classmethod
def set_forward(cls):
forward_source = inspect.getsource(LlamaForCausalLM.forward)
forward_source, _ = detab_code(forward_source)
cls.forward = types.MethodType(
compile(forward_source, "<forward>", "exec"), cls
)
@classmethod
def enable_act_offloading(cls):
forward_source = inspect.getsource(cls.forward)
forward_source = forward_source.replace(
HF_MODEL_OUTPUTS, PATCHED_HF_MODEL_OUTPUTS
)
forward_source, _ = detab_code(forward_source)
# replace forward method with patched version
cls.forward = types.MethodType(
compile(forward_source, "<llama_forward_w_act_offloading>", "exec"), cls
)
cls.act_offloading_ctx_manager = OffloadActivations()
@classmethod
def enable_liger_fce(cls, enable_act_offloading=True):
from liger_kernel.transformers.model.llama import (
lce_forward as llama_lce_forward,
)
if enable_act_offloading:
lce_source = inspect.getsource(llama_lce_forward)
lce_source = lce_source.replace(LCE_MODEL_OUTPUTS, PATCHED_LCE_OUTPUTS)
# replace forward method with patched version
cls.forward = types.MethodType(
compile(lce_source, "<llama_lce_forward_w_act_offloading>", "exec"),
cls,
)
else:
cls.forward = types.methodType(llama_lce_forward, cls)
@classmethod
def patch_hf_ga(cls):
# bugfix patch for gradient accumulation
forward_source = inspect.getsource(cls.forward)
forward_source = forward_source.replace(
HF_GA_FORWARD_1, PATCHED_HF_GA_FORWARD_1
)
forward_source = forward_source.replace(
HF_GA_FORWARD_2, PATCHED_HF_GA_FORWARD_2
)
forward_source, _ = detab_code(forward_source)
# replace forward method with patched version
cls.forward = types.MethodType(
compile(forward_source, "<llama_forward_ga_fix>", "exec"), cls
)
def replace_auto_model():
from transformers import LlamaConfig
from transformers.models.auto import MODEL_FOR_CAUSAL_LM_MAPPING
MODEL_FOR_CAUSAL_LM_MAPPING[LlamaConfig] = AxolotlLlamaForCausalLM
AxolotlLlamaForCausalLM.set_forward()
return AxolotlLlamaForCausalLM

View File

@@ -679,7 +679,6 @@ class AxolotlInputConfig(
default=False
)
gradient_checkpointing_kwargs: Optional[Dict[str, Any]] = None
activation_offloading: Optional[bool] = None
unfrozen_parameters: Optional[List[str]] = None

View File

@@ -380,15 +380,6 @@ class ModelLoader:
plugin_manager = PluginManager.get_instance()
plugin_manager.pre_model_load(self.cfg)
if self.cfg.model_config_type == "llama":
from axolotl.monkeypatch.models.llama.modeling_llama import replace_auto_model
AxolotlLlamaForCausalLM = replace_auto_model()
AxolotlLlamaForCausalLM.patch_hf_ga()
if self.cfg.activation_offloading:
AxolotlLlamaForCausalLM.enable_act_offloading()
if self.cfg.fsdp:
from axolotl.monkeypatch.trainer_fsdp_optim import (
patch_training_loop_for_fsdp,
@@ -1192,8 +1183,6 @@ class ModelLoader:
self.apply_lora_patch()
# self.apply_patches_to_model()
for _ in range(3):
gc.collect()
torch.cuda.empty_cache()