Wing Lian
cae5cebb59
xformers attention with packing ( #2619 )
...
* xformers attention with packing
* wire up the patch
* fix xformers + packing validation
* fix warning
* reorder the packing check
* fix fp16 / bf16 reset when using fp16 with bf16 auto
* fix seq lens calc to drop hanging sequences
* handle xformers patch for inference too
* fix batch size setter
* fix xformers inference
* add colab callback to fix inference post train
* PR feedback
2025-05-07 16:10:16 -04:00
Wing Lian
fcbd7477d0
Multipack parallel bin packing ( #2631 )
...
* improve readability of multipack sampler
* parallel bin packing
fix error with lambda and pickling
make sure things are in float instead of np.float
* annotations and comments update
* support for configurable group and bin size for sample packing
* fix missing map back to original indices
2025-05-07 16:10:15 -04:00
Wing Lian
038db85a40
allow plugins to return their own dataset ( #2617 ) [skip ci]
...
* allow plugins to return their own dataset
* add post_trainer_create and wire up
* add hook check
* address PR feedback:
* remove annotation causing circular import
2025-05-07 16:10:15 -04:00
NanoCode012
680dcc5a4d
feat(doc): add split_thinking docs ( #2613 ) [skip ci]
...
* feat(doc): add split_thinking docs
* fix: link config.qmd to conversation.qmd for split_thinking example
* update thinking => reasoning_content in messages format
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-07 16:10:15 -04:00
Wing Lian
fed5ca8254
bump liger dep to 0.5.9 ( #2640 ) [skip ci]
...
* bump liger dep to 0.5.9
* also upgrade vllm to post1, and datasets to 3.5.1
2025-05-07 16:10:15 -04:00
mhenrichsen
7a2d017c88
Update lr_scheduler options in config.qmd to include additional scheduling strategies for improved training flexibility. ( #2636 ) [skip ci]
2025-05-07 16:10:15 -04:00
Wing Lian
8c0303aa5e
Print axolotl art if train is called outside of cli: ( #2627 ) [skip ci]
2025-05-07 16:10:14 -04:00
Wing Lian
5d61169f7c
fix dpo eval override to call grandparent instead of the broken super ( #2628 ) [skip ci]
2025-05-07 16:10:14 -04:00
Wing Lian
e1586f7919
make sure gc_steps is used for all trainers ( #2638 )
2025-05-07 16:10:14 -04:00
Wing Lian
e4bf3ffb17
repop cache ( #2639 )
...
* repop cache
* pre-cache as a step
* fix the name
* add reason for pytest skipif
* restore pytorch matrix
* remove max-parallel now that we've optimized this a bit
2025-05-07 16:10:14 -04:00
mhenrichsen
30150fe1e1
Adds example for training a TTS model on top of a LLM. ( #2614 )
...
* Adds example for training a TTS model on top of a LLM.
* Update examples/orpheus/finetune.yml
Co-authored-by: NanoCode012 <nano@axolotl.ai >
* Update examples/orpheus/finetune.yml
Co-authored-by: NanoCode012 <nano@axolotl.ai >
* Update README.md to clarify GPU requirements for finetuning Orpheus TTS model
* Update finetune.yml to use the new base model canopylabs/orpheus-3b-0.1-pretrained
* Update finetune.yml and README.md for consistency and clarity
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-05-07 16:10:14 -04:00
Emmanuel Ferdman
7f7d7ade2e
Fix logging deprecation warnings ( #2623 )
...
Signed-off-by: Emmanuel Ferdman <emmanuelferdman@gmail.com >
2025-05-07 16:10:14 -04:00
Wing Lian
776cf70fe4
include multipack support for qwen3 family ( #2622 )
2025-05-07 16:10:14 -04:00
Wing Lian
8730951aba
setup hf transfer too and fix auto bf16 when fp16 enabled ( #2620 ) [skip ci]
2025-05-07 16:10:13 -04:00
Wing Lian
e72c11ad55
qwen3 and qwen3_moe support for liger kernels ( #2612 )
...
* qwen3 and qwen3_moe support for liger kernels
* fix moe module path
* fix: qwen3 liger input args and mlp
* fix: qwen3 input args and output class
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-05-07 16:10:13 -04:00
aitechguy
1a7978b960
remove keys to incoporate changes for the trl update ( #2616 )
2025-05-07 16:10:13 -04:00
Wing Lian
60b0d14f1d
automatically set pad_to_sequence_len when use packing ( #2607 )
...
* automatically set pad_to_sequence_len when use packing
* update tests
2025-05-07 16:10:13 -04:00
NanoCode012
a7a40378f5
fix: run preview-docs only when md/qmd changes ( #2606 )
...
* fix: run preview-docs only when md/qmd changes
* feat: add quarto yaml based on PR feedback
2025-05-07 16:10:13 -04:00
Wing Lian
b50d35bec9
Logging config for colab ( #2611 )
...
* only configure logging on cli to play nicely with colab
* allow reloading the config on the fly from a dict
* make sure to use dict for yaml
* reuse existing function for load
* make cli args optional
* mps fix and respect max_steps
2025-05-07 16:10:13 -04:00
Wing Lian
bc6dfa6899
add missing __init__ for lr monkeypatch fix ( #2609 )
2025-05-07 16:10:13 -04:00
Dhruv Mullick
9d6e8af622
Add num_completions_to_print for trl and grpo ( #2604 )
2025-05-07 16:10:12 -04:00
Wing Lian
17b441248c
use latest hf-xet and don't install vllm for torch 2.7.0 ( #2603 )
...
* use latest hf-xet and don't install vllm for torch 2.7.0
* fix runpod hub tests
2025-05-07 16:10:12 -04:00
Wing Lian
d49a4268b8
additional args for grpo config/trainer ( #2598 )
2025-05-07 16:10:12 -04:00
Wing Lian
1d6e931115
replace zero_only with simpler if statement ( #2592 )
2025-05-07 16:10:12 -04:00
Wing Lian
ff106ace44
ensure we pass axolotl extras to the Dockerfile so vllm is included in shipped images ( #2599 )
2025-05-07 16:10:12 -04:00
Wing Lian
24907533d1
don't automatically enable lora kernels for RL training ( #2600 )
2025-05-07 16:10:12 -04:00
Wing Lian
0e9d816d2e
only import vllm serve cli if its being called ( #2597 ) [skip ci]
2025-05-07 16:10:12 -04:00
Wing Lian
72f142186a
Handle other reasoning trace dataset formats ( #2591 )
...
* Handle other reasoning trace dataset formats
* rename var to improve readability
* chore: refactor with comments
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-05-07 16:10:11 -04:00
Wing Lian
87726322bf
upload the deepspeed json to wandb ( #2593 ) [skip ci]
2025-05-07 16:10:11 -04:00
NanoCode012
ae8ae7534c
feat: add qwen3 moe block for ds3 ( #2596 ) [skip ci]
2025-05-07 16:10:11 -04:00
Wing Lian
ee00142cb5
patch to convert LR from tensor to float when using DS ( #2595 ) [skip ci]
2025-05-07 16:10:11 -04:00
Aleksandr Dremov
097e7e3b5b
Plugins create_lr_scheduler support ( #2584 )
...
* lr_scheduler support
* fix
* Update scheduler.py
* Update scheduler.py
* cfg handling
* black
* remove debug
* remove adding the axolotl cfg to the scheduler mixin
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-07 16:10:11 -04:00
Dan Saunders
c714958181
auto-enable lora kernels where possible ( #2589 )
...
* auto-enable lora kernels where possible
* test
* revert change to example yaml
* naming
* remove print
* slight logic change
2025-05-07 16:10:11 -04:00
NanoCode012
4402c293dc
fix(doc): key used to point to url in multimodal doc ( #2575 ) [skip ci]
2025-05-07 16:10:10 -04:00
Wing Lian
0d71f787a3
bump vllm==0.8.5 for qwen3 support ( #2583 ) [skip ci]
2025-05-07 16:10:10 -04:00
Wing Lian
c337ca0872
support for qwen3 with lora kernels ( #2588 )
...
* support for qwen3 with lora kernels
* fix patch
* typo
2025-05-07 16:10:10 -04:00
Dan Saunders
f04f7cf5ad
Fix eval + add smoke test ( #2586 )
...
* fix evaluate CLI
* add smoke test
* fix naming
* lint
2025-05-07 16:10:10 -04:00
Wing Lian
c64a951bc9
set config on the PluginManager for callback access ( #2587 )
2025-05-07 16:10:10 -04:00
Wing Lian
fc88cc56cb
Post release fixes ( #2581 )
...
* fix missing kwarg on child
* make the runpod test shorter
* update docs
* rename runpod test json file
* typing fixes and ordering of doc
2025-05-07 16:10:10 -04:00
Wing Lian
e85cbb8645
remove torch 2.4.1 CI as part of support deprecation ( #2582 )
2025-05-07 16:10:10 -04:00
Wing Lian
14d670dbf0
v0.9.0 release ( #2578 )
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 124, 12.4.1, true, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 126, 12.6.3, 3.11, 2.7.0) (push) Has been cancelled
publish pypi / Create Release (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, true, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 126, 12.6.3, 3.11, 2.7.0) (push) Has been cancelled
ci-cd / build-axolotl-cloud-no-tmux (<nil>, 124, 12.4.1, 3.11, 2.6.0) (push) Has been cancelled
publish pypi / Upload release to PyPI (push) Has been cancelled
v0.9.0
2025-04-28 18:23:17 -04:00
Wing Lian
2d77165dc0
automatically split out reasoning trace from dataset ( #2579 )
...
* automatically split out reasoning trace from dataset
* chore: lint
* fix import
2025-04-28 18:23:03 -04:00
Wing Lian
63b17e3109
chat template and example for qwen3 ( #2577 )
2025-04-28 15:09:41 -04:00
NanoCode012
1178a15ede
Feat: Add qwen3 and CCE for qwen family ( #2518 )
2025-04-28 12:18:46 -04:00
Wing Lian
c513487d1a
support val_set_size for splitting test split from train with DPO ( #2572 )
2025-04-28 12:12:15 -04:00
Dan Saunders
dda95e6c40
add preview-docs workflow ( #2432 )
...
* add preview-docs workflow
* update preview-docs workflow
* use correct publish-dir
* install deps prior to docs build
* use correct publish-dir
* use quarto publish with netlify target
* adding _publish.yml
* fix
* fix
* fix
* remove unused file
* fix naming
---------
Co-authored-by: Dan Saunders <dan@axolotl.ai >
2025-04-28 11:20:46 -04:00
NanoCode012
7099343c56
feat: add eos_tokens and train_on_eot for chat_template EOT parsing ( #2364 )
...
* feat: add eos_tokens and train_on_eot for chat_template EOT parsing
* fix: comments
* chore: add some examples of tokens
* feat: add new potential errors for chat_template to faq
* feat: add examples for EOT handling
* fix: change error to warning for missing EOS
* fix: warning typo
* feat: add tests for eot token handling
* fix: remove broken caplog capture in test
* fix: chattemplate strategy with kd missing eot changes
2025-04-28 10:11:20 -04:00
Wing Lian
5000cb3fe7
grab sys prompt too from dataset ( #2397 ) [skip ci]
...
* grab sys prompt too from dataset
* chore: add field_system to docs
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-04-28 10:11:06 -04:00
divyanshuaggarwal
170cdb5be9
Add Post_model_load, post_lora_load, post_train, post_train_unload function calls ( #2539 )
...
* Update train.py
add post_model_load and post_lora_load model calss.
* Update train.py
add post_train and post_train_unload function calls
* Update train.py
* Update base.py
* Update train.py
* chore: lint
* clarify plugin hooks
* Update src/axolotl/integrations/base.py
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
* Update src/axolotl/utils/models.py
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
* Update src/axolotl/utils/models.py
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
* Update src/axolotl/integrations/base.py
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
* Update models.py
* Update models.py
* remove extra call to post_model_load
* chore: lint
* add test for hooks and gc trainer
* disable duplicated code check for test
* fix the path and add better handling
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
2025-04-28 10:10:28 -04:00
Ezekiel Wotring
5d182a1056
Add runpod sls handler ( #2530 ) [skip ci]
...
* Add runpod sls handler
* remove LICENSE and fix README
* chore: lint
* use axolotl cloud image as base and various fixes
* fix: trim allowed cuda versions
* restore dockerfile
* chore: update title
* use axolotl cloud image
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-04-28 10:08:32 -04:00