Wing Lian
c7f38ba96b
fix seq lens calc to drop hanging sequences
2025-05-03 21:56:45 -04:00
Wing Lian
372fd08548
fix fp16 / bf16 reset when using fp16 with bf16 auto
2025-05-03 21:56:39 -04:00
Wing Lian
52cab2aa5b
refactor so we can add test
2025-05-03 21:55:34 -04:00
Wing Lian
bed8f354a5
reorder the packing check
2025-05-03 15:38:29 -04:00
Wing Lian
f301a165c3
fix xformers + packing validation
2025-05-03 15:00:33 -04:00
Wing Lian
2b3a09aeae
wire up the patch
2025-05-03 15:00:29 -04:00
Wing Lian
648780de51
xformers attention with packing
2025-05-03 14:59:49 -04:00
Wing Lian
ecc2388274
chunked cross entropy loss
2025-05-03 14:59:43 -04:00
Wing Lian
ebf724a9d9
fix import
2025-05-03 12:03:15 -04:00
Wing Lian
99095573c3
add tabs back to code check
2025-05-03 12:03:15 -04:00
Wing Lian
140083a828
patch peft to not upcast everything
2025-05-03 12:03:15 -04:00
Wing Lian
37c27aedc1
fsdp embeddings should be float32 per comment
2025-05-03 12:03:15 -04:00
Wing Lian
ed922796b7
include multipack support for qwen3 family ( #2622 )
2025-05-03 12:02:39 -04:00
Wing Lian
3dd9c3bf3f
setup hf transfer too and fix auto bf16 when fp16 enabled ( #2620 ) [skip ci]
2025-05-03 12:02:26 -04:00
Wing Lian
0ba7d362fa
qwen3 and qwen3_moe support for liger kernels ( #2612 )
...
* qwen3 and qwen3_moe support for liger kernels
* fix moe module path
* fix: qwen3 liger input args and mlp
* fix: qwen3 input args and output class
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-05-02 09:29:55 -04:00
aitechguy
e4f73bc98e
remove keys to incoporate changes for the trl update ( #2616 )
2025-05-02 08:47:42 -04:00
Wing Lian
bcb59c70e2
automatically set pad_to_sequence_len when use packing ( #2607 )
...
* automatically set pad_to_sequence_len when use packing
* update tests
2025-05-01 13:24:38 -04:00
NanoCode012
6a3e6f8c53
fix: run preview-docs only when md/qmd changes ( #2606 )
...
* fix: run preview-docs only when md/qmd changes
* feat: add quarto yaml based on PR feedback
2025-05-01 13:21:28 -04:00
Wing Lian
fee3c13bb5
Logging config for colab ( #2611 )
...
* only configure logging on cli to play nicely with colab
* allow reloading the config on the fly from a dict
* make sure to use dict for yaml
* reuse existing function for load
* make cli args optional
* mps fix and respect max_steps
2025-05-01 12:58:00 -04:00
Rahul Tuli
996fc124e5
Add: Sparse Finetuning Integration with llmcompressor ( #2479 )
...
* Add: SFTPlugin with llmcompressor
* Update: review comments!
* Add:llmcompressor instalable
* pre commit hooks
* Use: warning over warn
* Revert: TODO's
* Update llmcompressor version to latest
* Apply suggestions from @markurtz
Co-authored-by: Mark Kurtz <mark.j.kurtz@gmail.com >
* Address review comments from @markurtz
* Add: llcompressor installable
* Rename: sft.yaml to sparse-finetuning.yaml
* Use: absolute import
* Update model config
* Move: LLMCompressorPlugin into it's own submodule
* Add: `llm_compressor` integration documentation
* Rebase and updates!
* Tests, Style, Updates
* Add: .qmd file
* Address Review Comments:
* deleted redundant docs/llm_compressor.qmd
* incorporated feedback in integration README.md
* added llmcompressor integration to docs/custom_integrations.qmd
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* Add: line about further optimizations using llmcompressor
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* Apply patch from @winglian
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* Fix: Test
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* additional fixes for docker and saving compressed
* split llmcompressor from vllm checks
* Reset session between tests
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
* move decorator to test method instead of class
* make sure to reset the session after each test
* move import of llmcompressor to reset session inside test
---------
Signed-off-by: Rahul Tuli <rtuli@redhat.com >
Co-authored-by: Mark Kurtz <mark.j.kurtz@gmail.com >
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-01 12:25:16 -04:00
Wing Lian
e963990ad7
add missing __init__ for lr monkeypatch fix ( #2609 )
2025-05-01 09:41:32 -04:00
Dhruv Mullick
c3f2b1c5c2
Add num_completions_to_print for trl and grpo ( #2604 )
2025-04-30 21:00:30 -04:00
Wing Lian
6ba5c0ed2c
use latest hf-xet and don't install vllm for torch 2.7.0 ( #2603 )
...
* use latest hf-xet and don't install vllm for torch 2.7.0
* fix runpod hub tests
2025-04-30 18:27:39 -04:00
Wing Lian
24ff5f53f8
additional args for grpo config/trainer ( #2598 )
2025-04-30 13:11:12 -04:00
Wing Lian
5e949eaa07
replace zero_only with simpler if statement ( #2592 )
2025-04-30 13:11:03 -04:00
Wing Lian
89ca14d9a0
ensure we pass axolotl extras to the Dockerfile so vllm is included in shipped images ( #2599 )
2025-04-30 11:35:45 -04:00
Wing Lian
8446b4ad28
don't automatically enable lora kernels for RL training ( #2600 )
2025-04-30 11:06:50 -04:00
Wing Lian
fc79606b6d
only import vllm serve cli if its being called ( #2597 ) [skip ci]
2025-04-30 09:11:25 -04:00
Wing Lian
baeb00231b
Handle other reasoning trace dataset formats ( #2591 )
...
* Handle other reasoning trace dataset formats
* rename var to improve readability
* chore: refactor with comments
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-04-30 03:32:55 -04:00
Wing Lian
2413688b08
upload the deepspeed json to wandb ( #2593 ) [skip ci]
2025-04-30 03:32:44 -04:00
NanoCode012
5bb1f3da56
feat: add qwen3 moe block for ds3 ( #2596 ) [skip ci]
2025-04-30 03:32:23 -04:00
Wing Lian
a21b9cc472
patch to convert LR from tensor to float when using DS ( #2595 ) [skip ci]
2025-04-30 03:31:57 -04:00
Aleksandr Dremov
41a1ec0c95
Plugins create_lr_scheduler support ( #2584 )
...
* lr_scheduler support
* fix
* Update scheduler.py
* Update scheduler.py
* cfg handling
* black
* remove debug
* remove adding the axolotl cfg to the scheduler mixin
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-04-29 17:08:30 -04:00
Dan Saunders
ecac731922
auto-enable lora kernels where possible ( #2589 )
...
* auto-enable lora kernels where possible
* test
* revert change to example yaml
* naming
* remove print
* slight logic change
2025-04-29 16:18:49 -04:00
NanoCode012
742fef4200
fix(doc): key used to point to url in multimodal doc ( #2575 ) [skip ci]
2025-04-29 15:10:59 -04:00
Wing Lian
a39caf8824
bump vllm==0.8.5 for qwen3 support ( #2583 ) [skip ci]
2025-04-29 15:10:40 -04:00
Wing Lian
07e4f2e25b
support for qwen3 with lora kernels ( #2588 )
...
* support for qwen3 with lora kernels
* fix patch
* typo
2025-04-29 15:02:49 -04:00
Dan Saunders
c7d07de6b4
Fix eval + add smoke test ( #2586 )
...
* fix evaluate CLI
* add smoke test
* fix naming
* lint
2025-04-29 12:58:54 -04:00
Wing Lian
6565ae85d8
set config on the PluginManager for callback access ( #2587 )
2025-04-29 12:05:44 -04:00
Wing Lian
80b4edb4a7
Post release fixes ( #2581 )
...
* fix missing kwarg on child
* make the runpod test shorter
* update docs
* rename runpod test json file
* typing fixes and ordering of doc
2025-04-29 10:01:38 -04:00
Wing Lian
fedbcc0254
remove torch 2.4.1 CI as part of support deprecation ( #2582 )
2025-04-29 08:28:32 -04:00
Wing Lian
8175896ada
add dev tag for v0.10.0.dev0 ( #2580 )
2025-04-28 20:30:14 -04:00
Wing Lian
14d670dbf0
v0.9.0 release ( #2578 )
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 124, 12.4.1, true, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 126, 12.6.3, 3.11, 2.7.0) (push) Has been cancelled
publish pypi / Create Release (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, true, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 126, 12.6.3, 3.11, 2.7.0) (push) Has been cancelled
ci-cd / build-axolotl-cloud-no-tmux (<nil>, 124, 12.4.1, 3.11, 2.6.0) (push) Has been cancelled
publish pypi / Upload release to PyPI (push) Has been cancelled
v0.9.0
2025-04-28 18:23:17 -04:00
Wing Lian
2d77165dc0
automatically split out reasoning trace from dataset ( #2579 )
...
* automatically split out reasoning trace from dataset
* chore: lint
* fix import
2025-04-28 18:23:03 -04:00
Wing Lian
63b17e3109
chat template and example for qwen3 ( #2577 )
2025-04-28 15:09:41 -04:00
NanoCode012
1178a15ede
Feat: Add qwen3 and CCE for qwen family ( #2518 )
2025-04-28 12:18:46 -04:00
Wing Lian
c513487d1a
support val_set_size for splitting test split from train with DPO ( #2572 )
2025-04-28 12:12:15 -04:00
Dan Saunders
dda95e6c40
add preview-docs workflow ( #2432 )
...
* add preview-docs workflow
* update preview-docs workflow
* use correct publish-dir
* install deps prior to docs build
* use correct publish-dir
* use quarto publish with netlify target
* adding _publish.yml
* fix
* fix
* fix
* remove unused file
* fix naming
---------
Co-authored-by: Dan Saunders <dan@axolotl.ai >
2025-04-28 11:20:46 -04:00
NanoCode012
7099343c56
feat: add eos_tokens and train_on_eot for chat_template EOT parsing ( #2364 )
...
* feat: add eos_tokens and train_on_eot for chat_template EOT parsing
* fix: comments
* chore: add some examples of tokens
* feat: add new potential errors for chat_template to faq
* feat: add examples for EOT handling
* fix: change error to warning for missing EOS
* fix: warning typo
* feat: add tests for eot token handling
* fix: remove broken caplog capture in test
* fix: chattemplate strategy with kd missing eot changes
2025-04-28 10:11:20 -04:00
Wing Lian
5000cb3fe7
grab sys prompt too from dataset ( #2397 ) [skip ci]
...
* grab sys prompt too from dataset
* chore: add field_system to docs
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-04-28 10:11:06 -04:00