xzuyn
f943306263
Add CAME Optimizer ( #2385 )
2025-05-07 16:10:17 -04:00
Wing Lian
493910559a
Configurable embeddings upcast ( #2621 )
...
* fsdp embeddings should be float32 per comment
* patch peft to not upcast everything
* add tabs back to code check
* fix import
* add configurable option and fix check
* add check for dtypes
* move embeddings test to patch dir
* fix test
* fix comment and logic
2025-05-07 16:10:16 -04:00
Wing Lian
cae5cebb59
xformers attention with packing ( #2619 )
...
* xformers attention with packing
* wire up the patch
* fix xformers + packing validation
* fix warning
* reorder the packing check
* fix fp16 / bf16 reset when using fp16 with bf16 auto
* fix seq lens calc to drop hanging sequences
* handle xformers patch for inference too
* fix batch size setter
* fix xformers inference
* add colab callback to fix inference post train
* PR feedback
2025-05-07 16:10:16 -04:00
NanoCode012
680dcc5a4d
feat(doc): add split_thinking docs ( #2613 ) [skip ci]
...
* feat(doc): add split_thinking docs
* fix: link config.qmd to conversation.qmd for split_thinking example
* update thinking => reasoning_content in messages format
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-05-07 16:10:15 -04:00
mhenrichsen
7a2d017c88
Update lr_scheduler options in config.qmd to include additional scheduling strategies for improved training flexibility. ( #2636 ) [skip ci]
2025-05-07 16:10:15 -04:00
NanoCode012
4402c293dc
fix(doc): key used to point to url in multimodal doc ( #2575 ) [skip ci]
2025-05-07 16:10:10 -04:00
Wing Lian
fc88cc56cb
Post release fixes ( #2581 )
...
* fix missing kwarg on child
* make the runpod test shorter
* update docs
* rename runpod test json file
* typing fixes and ordering of doc
2025-05-07 16:10:10 -04:00
NanoCode012
7099343c56
feat: add eos_tokens and train_on_eot for chat_template EOT parsing ( #2364 )
...
* feat: add eos_tokens and train_on_eot for chat_template EOT parsing
* fix: comments
* chore: add some examples of tokens
* feat: add new potential errors for chat_template to faq
* feat: add examples for EOT handling
* fix: change error to warning for missing EOS
* fix: warning typo
* feat: add tests for eot token handling
* fix: remove broken caplog capture in test
* fix: chattemplate strategy with kd missing eot changes
2025-04-28 10:11:20 -04:00
Wing Lian
5000cb3fe7
grab sys prompt too from dataset ( #2397 ) [skip ci]
...
* grab sys prompt too from dataset
* chore: add field_system to docs
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-04-28 10:11:06 -04:00
NanoCode012
f1df73a798
fix(doc): clarify vllm usage with grpo ( #2573 ) [skip ci]
...
* fix(doc): clarify vllm usage with grpo
* nit
Co-authored-by: salman <salman.mohammadi@outlook.com >
* Update docs/rlhf.qmd
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-04-28 10:07:45 -04:00
NanoCode012
9eba0ad118
chore(doc): update docker tags on doc ( #2559 ) [skip ci]
2025-04-25 17:14:48 -04:00
NanoCode012
85053f4bd4
Fix(doc): add delinearize instruction ( #2545 )
...
* fix: mention to install pytorch before axolotl
* feat(doc): include instruction to delinearize
* fix: update instruction for delinearize with adapter
2025-04-24 01:03:43 -04:00
Dan Saunders
b8c633aa97
batch api HF adapter for ring-flash-attn; cleanup and improvements ( #2520 )
...
* batch api HF adapter for ring-flash-attn; cleanup and improvements
* update
* adding all batch ring-flash-attn methods via single adapter
* removing pad_to_sequence_len=False for now
* fix
* updating docs to include batch SP
* review comments
* fixes for batch API funcs, simplify
* fixes
* fix
* updates
* add batch_zigzag smoke test
2025-04-16 13:50:48 -04:00
NanoCode012
51267ded04
chore: update doc links ( #2509 )
...
* chore: update doc links
* fix: address pr feedback
2025-04-11 09:53:18 -04:00
NanoCode012
756a0559c1
feat(doc): explain deepspeed configs ( #2514 ) [skip ci]
...
* feat(doc): explain deepspeed configs
* fix: add fetch configs
2025-04-11 09:52:43 -04:00
Sung Ching Liu
22c562533d
Update rlhf.qmd ( #2519 )
...
Fix typo in command that spawns a vllm server, should be `axolotl vllm-serve` not `axolotl vllm_serve`
2025-04-10 11:33:09 -04:00
NanoCode012
9b89591ead
Feat: Add doc on loading datasets and support for Azure/OCI ( #2482 )
...
* fix: remove unused config
* feat: add doc on dataset loading
* feat: enable azure and oci remote file system
* feat: add adlfs and ocifs to requirements
* fix: add links between dataset formats and dataset loading
* fix: remove unused condition
* Revert "fix: remove unused condition"
This reverts commit 5fe13be73e .
2025-04-07 12:41:13 -04:00
NanoCode012
31498d0230
fix(doc): clarify roles mapping in chat_template ( #2490 ) [skip ci]
2025-04-07 12:40:32 -04:00
NanoCode012
e0e5d9b1d6
feat: add llama4 multimodal ( #2499 )
...
* feat: add llama4 multimodal
* feat: add torchvision to base docker
* just use latest torchvision
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-04-07 10:49:29 -04:00
NanoCode012
adb593abac
fix: document offload gradient_checkpointing option ( #2475 )
2025-04-02 09:35:42 -04:00
NanoCode012
45bf634d17
feat: add support for multimodal in lora kernels ( #2472 ) [skip ci]
...
* feat: add support for multimodal in lora kernels
* fix: improve multimodal checks
* fix: add fallback for model config
* chor: add gemma3 to docs
2025-04-02 09:33:46 -04:00
NanoCode012
f4ae8816bb
Fix: remove the numerous sequential log ( #2461 )
...
* fix: remove sequential logs
* feat(doc): add for sample pack sequentially and curriculum sampling
2025-04-01 09:20:00 -04:00
NanoCode012
9b95e06cbb
Fix(doc): Minor doc changes for peft and modal ( #2462 ) [skip ci]
...
* fix(doc): document peft configs
* fix(doc): explain modal env vs secrets difference
* fix(doc): clarify evaluate vs lm-eval
* fix: clarify what is performance
2025-04-01 08:48:36 -04:00
NanoCode012
7acf93b59f
Fix(doc): Clarify doc on attention configs and missing pad_token ( #2455 ) [skip ci]
...
* fix: clarify input type
* fix: handling of error message if data_files not available
* fix: clarify attention handling
* fix: add doc on missing pad token
2025-03-31 15:47:28 -04:00
Wing Lian
b6fc46ada8
Updates for trl 0.16.0 - mostly for GRPO ( #2437 ) [skip ci]
...
* add grpo scale_rewards config for trl#3135
* options to connect to vllm server directly w grpo trl#3094
* temperature support trl#3029
* sampling/generation kwargs for grpo trl#2989
* make vllm_enable_prefix_caching a config param trl#2900
* grpo multi-step optimizeations trl#2899
* remove overrides for grpo trainer
* bump trl to 0.16.0
* add cli to start vllm-serve via trl
* call the python module directly
* update to use vllm with 2.6.0 too now and call trl vllm serve from module
* vllm 0.8.1
* use python3
* use sys.executable
* remove context and wait for start
* fixes to make it actually work
* fixes so the grpo tests pass with new vllm paradigm
* explicit host/port and check in start vllm
* make sure that vllm doesn't hang by setting quiet so outouts go to dev null
* also bump bnb to latest release
* add option for wait from cli and nccl debugging for ci
* grpo + vllm test on separate devices for now
* make sure grpo + vllm tests runs single worker since pynccl comms would conflict
* fix cli
* remove wait and add caching for argilla dataset
* refactoring configs
* chore: lint
* add vllm config
* fixup vllm grpo args
* fix one more incorrect schema/config path
* fix another vlllm reference and increase timeout
* make the tests run a bit faster
* change mbsz back so it is correct for grpo
* another change mbsz back so it is correct for grpo
* fixing cli args
* nits
* adding docs
* docs
* include tensor parallel size for vllm in pydantic schema
* moving start_vllm, more docs
* limit output len for grpo vllm
* vllm enable_prefix_caching isn't a bool cli arg
* fix env ordering in tests and also use pid check when looking for vllm
---------
Co-authored-by: Salman Mohammadi <salman.mohammadi@outlook.com >
2025-03-31 15:47:11 -04:00
Dan Saunders
5410195e0b
Sequence parallelism quick follow-ups; remove ModelCallback ( #2450 )
...
* guard return if ring attn alrady registered
* add docs link, bits in multi-gpu docs, remove save model callback (subsumed by HF trainers)
* configurable heads_k_stride from ring-flash-attn hf adapter
2025-03-31 09:13:42 -04:00
NanoCode012
a7811ad4a0
fix(doc): document config required to run eval_causal_lm_metrics ( #2445 ) [skip ci]
2025-03-26 18:14:29 -04:00
NanoCode012
e2da821e67
chore: minor optim changes (add apollo, improve docs, remove lion-pytorch) ( #2444 )
...
* feat: add apollo-torch
* chore: update optimizer list
* fix: deleted accidental requirements file
* fix: remove mention of deprecated lion_pytorch
2025-03-26 18:14:07 -04:00
NanoCode012
2c34a4634e
feat: add CCE for gemma3, cohere, and cohere2 ( #2443 )
...
* feat: add CCE for gemma3 and cohere1/2
* fix: change from relative import to absolute
* feat: add multipack for cohere&cohere2
* chore: improve comments
* fix: add gemma3_text
* feat: add cohere2 example
* fix: cohere forward
* fix: patch for cohere2
* feat: add command r v01 qlora sample
* chore: lint
* feat: upgrade gemma3 and gemma2 patch to use logits_to_keep
* chore: lint
* fix: add deprecate_kwarg decorator
* fix: add cce for gemma3 conditionalgeneration
* fix: gemma3 patch to defer logits calculation
* fix: patch gemma3 if given as model
* fix: remove not working config
* fix: update comments to clarify changes
* feat(doc): add supported models to readme
* fix: address difference in our cohere patch
* feat: add mistral3
* feat: add gemma
* feat(doc): update README to include gemma and mistral3 in supported models
* fix: gemma patch
* fix: import
* fix: gemma patch to be standalone
* fix: gemma3 warn about not support final_logit_softcapping
* feat: add mllama CCE
* chore: add abbireviation to doc
* fix: remove unneeded gemma3 eager warning
* fix: save processor if available
* fix: enable save processor on merge
* fix: wrong env meaning
2025-03-26 18:13:51 -04:00
NanoCode012
a9b0733f2c
Feat: Rework multimodal support (mllama, llava, pixtral, qwen2, qwen25, gemma3, mistral3) ( #2435 )
2025-03-23 11:08:51 -04:00
NanoCode012
9f00465a5c
Feat: Add support for gemma3_text and add e2e for gemma2 ( #2406 )
2025-03-22 20:33:21 -04:00
Dan Saunders
23f0c51d88
Sequence parallelism ( #2412 )
...
* adding easy_context as integration for now
* progress on ring attn impl
* progress on ring attn impl
* cleanup
* remove errant file
* fix req
* removing unused code
* updates
* pytest
* update
* updates
* fixes
* precommit fixes
* working multi-group SP
* fixing sample packing
* remove debug logs and simplify
* eval dataloader and sampler changes
* removing some obvious comments
* update config.qmd and rename option
* scoping down problematic import
* another import scoping change
* pernicious Fire CLI bugfix
* isolate cli tests
* actually isolate CLI tests
* gracefully handle no ring-flash-attn
* fix
* fix
* move ring flash attn to extras with flash-attn (#2414 )
* removing flash-attn from requirements.txt (in setup.py extras already)
* rename file, delete another
* using field validator instead of model validator
* test fix
* sampler / dataloader refactor
* non-seq2se1 collator fix
* removing print statement
* bugfix
* add SP doc, review comments
* small changes
* review comments, docstrings
* refactors, SP mixin
* small updates
* fix tests
* precommit
* precommit
---------
Co-authored-by: Wing Lian <wing.lian@gmail.com >
Co-authored-by: Dan Saunders <dan@axolotl.ai >
2025-03-21 12:43:55 -04:00
Dan Saunders
113e9cd193
Autodoc generation with quartodoc ( #2419 )
...
* quartodoc integration
* quartodoc progress
* deletions
* Update docs/.gitignore to exclude auto-generated API documentation files
* Fix
* more autodoc progress
* moving reference up near the top of the sidebar
* fix broken link
* update to reflect recent changes
* pydantic models refactor + add to autodoc + fixes
* fix
* shrinking header sizes
* fix accidental change
* include quartodoc build step
* update pre-commit version
* update pylint
* pre-commit
---------
Co-authored-by: Dan Saunders <dan@axolotl.ai >
2025-03-21 12:26:47 -04:00
NanoCode012
61825a464a
chore(doc): add explanation on fsdp_transformer_layer_cls_to_wrap ( #2429 ) [skip ci]
2025-03-21 11:59:22 -04:00
NanoCode012
f8de8bb4f2
chore(doc): add instructions on adding custom integrations ( #2422 ) [skip ci]
...
* chore(doc): add instructions on adding custom integrations
* chore: add warning help
* feat: add note about integration path
* fix: adjust text per suggestion
2025-03-21 10:18:01 -04:00
hugo
8e604848a4
add run on novita ai ( #2421 ) [skip ci]
...
* add run on novita ai
* Revert "add run on novita ai"
This reverts commit 4d5df1ac6b .
* add run axolotl on novita ai
2025-03-21 10:17:47 -04:00
SicariusSicariiStuff
85147ec430
Update README.md ( #2360 )
...
* Update README.md
wheel is needed
* feat: add ninja, setuptools, packing to installation steps
* fix: add missing instruction
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-03-17 08:39:17 -04:00
NanoCode012
51cd409488
Feat: minor docs improvements for RLHF and faq on embeddings ( #2401 ) [skip ci]
...
* feat: add doc on shrink_embeddings and custom calling
* chore: rename inference doc
* fix: clarify same config is used for all cli
* chore: rearrange order inference qmd
* feat: add simpo to doc
* fix: update defaults
* feat: add rl configs to doc
* fix: ensure beta consistent with trl.beta
* fix: clarify about lora/fft
* chore: rename title
* chore: fix language
* feat: move config reference higher
* Update docs/getting-started.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
* Update docs/rlhf.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-17 08:39:04 -04:00
NanoCode012
7235123d44
chore(docs): add cookbook/blog link to docs ( #2410 ) [skip ci]
2025-03-17 08:38:19 -04:00
NanoCode012
46a045e528
chore(doc): add faq when having no default chat_template ( #2398 )
...
* chore(doc): add faq when having no default chat_template
* Update docs/dataset-formats/conversation.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
* Update docs/faq.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-10 16:25:50 +07:00
NanoCode012
3b477e08a0
feat(doc): add more info on RewardModel datasets ( #2391 )
...
* fix: reduce title size
* feat(doc): add rm dataset info
* Update docs/reward_modelling.qmd following suggestion
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-10 16:25:31 +07:00
NanoCode012
16dc6ee68d
refactor: trl grpo configs to have descriptions ( #2386 )
...
* refactor: trl grpo configs to have descriptions
* chore: caps
2025-03-07 08:58:53 -05:00
mhenrichsen
575e5f28ec
Update Tokenizer Overrides Handling in models.py ( #1549 )
...
* override special tokens mock code
* fix(doc): remove duplicate config
* feat: replace added_tokens in tokenizer and add test
* make sure to run tokenizer modification on rank 0 only
* use is local main process instead
* feat: rename config
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-03-05 11:15:12 -05:00
xzuyn
0134093acc
Add REX LR Scheduler ( #2380 )
...
* Update trainer_builder.py
* Update base.py
* Update __init__.py
* Update base.py
* Update base.py
* Update config.qmd
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* lint
* lint
* lint
* lint
* lint
* lint
* Update base.py
* Update base.py
* lint
* Update base.py
* Update base.py
* Move RexLR to `schedulers.py`
* Remove RexLR from `base.py`
* Fix tooltip formatting
* lint
* Create test_schedulers.py
* Use a default optimizer in test
* lint
* lint
* Add `warmup_steps` and `cosine_min_lr_ratio` to test
* lint
2025-03-05 10:26:11 -05:00
NanoCode012
c8191394e9
fix(doc): add missing low_cpu_mem_usage config to docs ( #2369 ) [skip ci]
2025-03-05 10:01:44 -05:00
NanoCode012
f18231c653
chore(doc): add clarification about mpi4py error on single gpu deepspeed ( #2383 ) [skip ci]
...
* chore(doc): add clarification about mpi4py error on single gpu deepspeed
* fix: lint
2025-03-05 10:01:28 -05:00
NanoCode012
9ed4f6b3aa
feat(doc): document drop_system_message and clarify limitation ( #2381 ) [skip ci]
2025-03-05 10:01:16 -05:00
NanoCode012
05dddfc41d
feat(doc): add docker images explanation ( #2379 ) [skip ci]
...
* feat(doc): add docker images explanation
* chore: add link to dockerhub
2025-03-05 10:01:00 -05:00
NanoCode012
75cbd15301
Fix(doc): address missing doc changes ( #2362 )
...
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 124, 12.4.1, true, 3.11, 2.5.1) (push) Has been cancelled
publish pypi / Create Release (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, true, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud-no-tmux (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
publish pypi / Upload release to PyPI (push) Has been cancelled
* fix: add multiple tips about eos_token masking
* fix: format dataset preprocessing doc
* Update docs/dataset-formats/conversation.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-02-25 13:50:02 -05:00
NanoCode012
2efe1b4c09
Feat(doc): Reorganize documentation, fix broken syntax, update notes ( #2348 )
...
* feat(doc): organize docs, add to menu bar, fix broken formatting
* feat: add link to custom integrations
* feat: update readme for integrations to include citations and repo link
* chore: update lm_eval info
* chore: use fullname
* Update docs/cli.qmd per suggestion
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
* feat: add sweep doc
* feat: add kd doc
* fix: remove toc
* fix: update deprecation
* feat: add more info about chat_template issues
* fix: heading level
* fix: shell->bash code block
* fix: ray link
* fix(doc): heading level, header links, formatting
* feat: add grpo docs
* feat: add style changes
* fix: wrong cli arg for lm-eval
* fix: remove old run method
* feat: load custom integration doc dynamically
* fix: remove old cli way
* fix: toc
* fix: minor formatting
---------
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
2025-02-25 16:09:37 +07:00