NanoCode012
fd271b2547
fix: consolidate handling of fp16, bf16, tf32 kwarg
2025-05-14 09:33:28 +07:00
NanoCode012
c268a0157a
feat: add report_to to set run name
2025-05-14 09:33:28 +07:00
NanoCode012
6317945b67
fix: refactor sft and rl trainer to set same base args
2025-05-14 09:32:46 +07:00
NanoCode012
86ba574698
feat: add num_proc and load from cache for rl mapping
2025-05-14 09:32:04 +07:00
Wing Lian
7fa1089cea
Atropos support ( #2666 ) [skip ci]
...
* allow peft+liger+grpo and custom vllm serve for atropos support
* set trainer class for RL
2025-05-13 08:30:58 -04:00
Dan Saunders
80304c26a7
SP GRPO support + batch SP fixes ( #2643 )
...
* ctx manager for SP
* updates
* update
* further simplifying
* simplifying
* simplifying
* reorg
* batch api HF adapter for ring-flash-attn; cleanup and improvements
* update
* adding all batch ring-flash-attn methods via single adapter
* fix
* fixes for batch API funcs, simplify
* fix
* grpo sp support
* progress
* stronger subclassing of TRL GRPO trainer; custom distributed sampler
* subclassing constructor
* progress
* finalizing SP + GRPO trainer
* minimize diffs to GRPO trainer
* remove (most of) the custom GRPO trainer logic
* debug
* debug
* update
* update
* update
* progress
* cleanup
* cleanup
* minor changes
* update
* update
* update
* small changes
* updates
* cleanup; torch.compile ring_flash_attn functions to prevent numerical instability; lint
* spacing
* cleanup; log in pydantic model config only on main process
* remove comment
* fix sp sampler, update to latest upstream code, doc
* add docs
* update quartodoc autodoc contents
* fix, simplifications
* fixes + simplifications
* review comments
* lint
* removing main process only logs in favor of #2608
* fixes, additional smoke test
* updatse
* more tests
* update
* fix grad accum bug (sort of)
* lint, tests
* todo
2025-05-12 17:52:40 -04:00
NanoCode012
67c4ea9c7c
fix: disable auto lora kernel if dropout nonzero ( #2655 ) [skip ci]
...
* fix: disable auto lora kernel if dropout nonzero
* Add comment from PR feedback
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-12 16:23:53 -04:00
Wing Lian
c7b6790614
Various fixes for CI, save_only_model for RL, prevent packing multiprocessing deadlocks ( #2661 )
...
* lean mistral ft tests, remove e2e torch 2.4.1 test
* make sure to pass save_only_model for RL
* more tests to make ci leaner, add cleanup to modal ci
* fix module for import in e2e tests
* use mp spawn to prevent deadlocks with packing
* make sure cleanup shell script is executable when cloned out
2025-05-12 10:51:18 -04:00
Dan Saunders
47e0e71bc8
don't sort multipack sampler ( #2657 )
...
* don't sort multipack sampler
* increased packing efficiency increases loss
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-09 20:28:58 -04:00
Wing Lian
0f3587174d
swap tinymodels that have safetensors for some ci tests ( #2641 )
2025-05-07 15:06:07 -04:00
xzuyn
25e6c5f9bd
Add CAME Optimizer ( #2385 )
2025-05-07 10:31:46 -04:00
NanoCode012
9daa04da90
Fix: improve error message on failed dataset load ( #2637 ) [skip ci]
...
* fix(log): clarify error on dataset loading failed
* fix: add path for easy tracking of broken config
* fix: improve error message based on pr feedback
2025-05-07 10:29:05 -04:00
Wing Lian
0d71b0aa5f
Configurable embeddings upcast ( #2621 )
...
* fsdp embeddings should be float32 per comment
* patch peft to not upcast everything
* add tabs back to code check
* fix import
* add configurable option and fix check
* add check for dtypes
* move embeddings test to patch dir
* fix test
* fix comment and logic
2025-05-06 23:40:44 -04:00
Eric Meier
63aaccf85b
Fix cut_cross_entropy plugin install ( #2642 ) [skip ci]
2025-05-06 22:56:00 -04:00
Wing Lian
ff0fe767c8
xformers attention with packing ( #2619 )
...
* xformers attention with packing
* wire up the patch
* fix xformers + packing validation
* fix warning
* reorder the packing check
* fix fp16 / bf16 reset when using fp16 with bf16 auto
* fix seq lens calc to drop hanging sequences
* handle xformers patch for inference too
* fix batch size setter
* fix xformers inference
* add colab callback to fix inference post train
* PR feedback
2025-05-06 22:49:22 -04:00
Wing Lian
8e4158cc0b
Multipack parallel bin packing ( #2631 )
...
* improve readability of multipack sampler
* parallel bin packing
fix error with lambda and pickling
make sure things are in float instead of np.float
* annotations and comments update
* support for configurable group and bin size for sample packing
* fix missing map back to original indices
2025-05-06 20:08:08 -04:00
Wing Lian
cd84325253
allow plugins to return their own dataset ( #2617 ) [skip ci]
...
* allow plugins to return their own dataset
* add post_trainer_create and wire up
* add hook check
* address PR feedback:
* remove annotation causing circular import
2025-05-06 20:05:51 -04:00
Wing Lian
b71c0e3447
Print axolotl art if train is called outside of cli: ( #2627 ) [skip ci]
2025-05-06 11:18:45 -04:00
Wing Lian
ddaebf8309
fix dpo eval override to call grandparent instead of the broken super ( #2628 ) [skip ci]
2025-05-06 11:18:25 -04:00
Wing Lian
679743087a
make sure gc_steps is used for all trainers ( #2638 )
2025-05-06 11:18:00 -04:00
Emmanuel Ferdman
54960d4de0
Fix logging deprecation warnings ( #2623 )
...
Signed-off-by: Emmanuel Ferdman <emmanuelferdman@gmail.com >
2025-05-04 08:22:45 -04:00
Wing Lian
ed922796b7
include multipack support for qwen3 family ( #2622 )
2025-05-03 12:02:39 -04:00
Wing Lian
3dd9c3bf3f
setup hf transfer too and fix auto bf16 when fp16 enabled ( #2620 ) [skip ci]
2025-05-03 12:02:26 -04:00
Wing Lian
0ba7d362fa
qwen3 and qwen3_moe support for liger kernels ( #2612 )
...
* qwen3 and qwen3_moe support for liger kernels
* fix moe module path
* fix: qwen3 liger input args and mlp
* fix: qwen3 input args and output class
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-05-02 09:29:55 -04:00
aitechguy
e4f73bc98e
remove keys to incoporate changes for the trl update ( #2616 )
2025-05-02 08:47:42 -04:00
Wing Lian
bcb59c70e2
automatically set pad_to_sequence_len when use packing ( #2607 )
...
* automatically set pad_to_sequence_len when use packing
* update tests
2025-05-01 13:24:38 -04:00
Wing Lian
fee3c13bb5
Logging config for colab ( #2611 )
...
* only configure logging on cli to play nicely with colab
* allow reloading the config on the fly from a dict
* make sure to use dict for yaml
* reuse existing function for load
* make cli args optional
* mps fix and respect max_steps
2025-05-01 12:58:00 -04:00
Rahul Tuli
996fc124e5
Add: Sparse Finetuning Integration with llmcompressor ( #2479 )
...
* Add: SFTPlugin with llmcompressor
* Update: review comments!
* Add:llmcompressor instalable
* pre commit hooks
* Use: warning over warn
* Revert: TODO's
* Update llmcompressor version to latest
* Apply suggestions from @markurtz
Co-authored-by: Mark Kurtz <mark.j.kurtz@gmail.com >
* Address review comments from @markurtz
* Add: llcompressor installable
* Rename: sft.yaml to sparse-finetuning.yaml
* Use: absolute import
* Update model config
* Move: LLMCompressorPlugin into it's own submodule
* Add: `llm_compressor` integration documentation
* Rebase and updates!
* Tests, Style, Updates
* Add: .qmd file
* Address Review Comments:
* deleted redundant docs/llm_compressor.qmd
* incorporated feedback in integration README.md
* added llmcompressor integration to docs/custom_integrations.qmd
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* Add: line about further optimizations using llmcompressor
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* Apply patch from @winglian
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* Fix: Test
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* additional fixes for docker and saving compressed
* split llmcompressor from vllm checks
* Reset session between tests
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* move decorator to test method instead of class
* make sure to reset the session after each test
* move import of llmcompressor to reset session inside test
---------
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
Co-authored-by: Mark Kurtz <mark.j.kurtz@gmail.com >
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-01 12:25:16 -04:00
Wing Lian
e963990ad7
add missing __init__ for lr monkeypatch fix ( #2609 )
2025-05-01 09:41:32 -04:00
Dhruv Mullick
c3f2b1c5c2
Add num_completions_to_print for trl and grpo ( #2604 )
2025-04-30 21:00:30 -04:00
Wing Lian
24ff5f53f8
additional args for grpo config/trainer ( #2598 )
2025-04-30 13:11:12 -04:00
Wing Lian
5e949eaa07
replace zero_only with simpler if statement ( #2592 )
2025-04-30 13:11:03 -04:00
Wing Lian
8446b4ad28
don't automatically enable lora kernels for RL training ( #2600 )
2025-04-30 11:06:50 -04:00
Wing Lian
fc79606b6d
only import vllm serve cli if its being called ( #2597 ) [skip ci]
2025-04-30 09:11:25 -04:00
Wing Lian
baeb00231b
Handle other reasoning trace dataset formats ( #2591 )
...
* Handle other reasoning trace dataset formats
* rename var to improve readability
* chore: refactor with comments
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-04-30 03:32:55 -04:00
Wing Lian
2413688b08
upload the deepspeed json to wandb ( #2593 ) [skip ci]
2025-04-30 03:32:44 -04:00
NanoCode012
5bb1f3da56
feat: add qwen3 moe block for ds3 ( #2596 ) [skip ci]
2025-04-30 03:32:23 -04:00
Wing Lian
a21b9cc472
patch to convert LR from tensor to float when using DS ( #2595 ) [skip ci]
2025-04-30 03:31:57 -04:00
Aleksandr Dremov
41a1ec0c95
Plugins create_lr_scheduler support ( #2584 )
...
* lr_scheduler support
* fix
* Update scheduler.py
* Update scheduler.py
* cfg handling
* black
* remove debug
* remove adding the axolotl cfg to the scheduler mixin
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-04-29 17:08:30 -04:00
Dan Saunders
ecac731922
auto-enable lora kernels where possible ( #2589 )
...
* auto-enable lora kernels where possible
* test
* revert change to example yaml
* naming
* remove print
* slight logic change
2025-04-29 16:18:49 -04:00
Wing Lian
07e4f2e25b
support for qwen3 with lora kernels ( #2588 )
...
* support for qwen3 with lora kernels
* fix patch
* typo
2025-04-29 15:02:49 -04:00
Dan Saunders
c7d07de6b4
Fix eval + add smoke test ( #2586 )
...
* fix evaluate CLI
* add smoke test
* fix naming
* lint
2025-04-29 12:58:54 -04:00
Wing Lian
6565ae85d8
set config on the PluginManager for callback access ( #2587 )
2025-04-29 12:05:44 -04:00
Wing Lian
80b4edb4a7
Post release fixes ( #2581 )
...
* fix missing kwarg on child
* make the runpod test shorter
* update docs
* rename runpod test json file
* typing fixes and ordering of doc
2025-04-29 10:01:38 -04:00
Wing Lian
8175896ada
add dev tag for v0.10.0.dev0 ( #2580 )
2025-04-28 20:30:14 -04:00
Wing Lian
14d670dbf0
v0.9.0 release ( #2578 )
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 124, 12.4.1, true, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 126, 12.6.3, 3.11, 2.7.0) (push) Has been cancelled
publish pypi / Create Release (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, true, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 126, 12.6.3, 3.11, 2.7.0) (push) Has been cancelled
ci-cd / build-axolotl-cloud-no-tmux (<nil>, 124, 12.4.1, 3.11, 2.6.0) (push) Has been cancelled
publish pypi / Upload release to PyPI (push) Has been cancelled
2025-04-28 18:23:17 -04:00
Wing Lian
2d77165dc0
automatically split out reasoning trace from dataset ( #2579 )
...
* automatically split out reasoning trace from dataset
* chore: lint
* fix import
2025-04-28 18:23:03 -04:00
Wing Lian
63b17e3109
chat template and example for qwen3 ( #2577 )
2025-04-28 15:09:41 -04:00
NanoCode012
1178a15ede
Feat: Add qwen3 and CCE for qwen family ( #2518 )
2025-04-28 12:18:46 -04:00
Wing Lian
c513487d1a
support val_set_size for splitting test split from train with DPO ( #2572 )
2025-04-28 12:12:15 -04:00