Dan Saunders
5b7e688fc5
fix broken link
2025-03-21 11:12:09 -04:00
Dan Saunders
5134aa66cd
moving reference up near the top of the sidebar
2025-03-21 11:12:09 -04:00
Dan Saunders
ba9a867adb
more autodoc progress
2025-03-21 11:12:09 -04:00
Dan Saunders
c618f42c39
Fix
2025-03-21 11:12:09 -04:00
Dan Saunders
fc1f985296
Update docs/.gitignore to exclude auto-generated API documentation files
2025-03-21 11:12:09 -04:00
Dan Saunders
a5e37f183c
deletions
2025-03-21 11:12:09 -04:00
Dan Saunders
e6a7bbe9ff
quartodoc progress
2025-03-21 11:12:09 -04:00
Dan Saunders
e4fd7aad0b
quartodoc integration
2025-03-21 11:12:09 -04:00
Dan Saunders
c907ac173e
adding pre-commit auto-update GH action and bumping plugin versions ( #2428 )
...
* adding pre-commit auto-update GH action and bumping plugin versions
* running updated pre-commit plugins
* sorry to revert, but pylint complained
* Update .pre-commit-config.yaml
Co-authored-by: Wing Lian <wing.lian@gmail.com >
---------
Co-authored-by: Dan Saunders <dan@axolotl.ai >
Co-authored-by: Wing Lian <wing.lian@gmail.com >
2025-03-21 11:02:43 -04:00
salman
187227d837
Fixing KTO+QLoRA+multi-GPU ( #2420 )
...
* WIP
* removing artifacts
* adding error
* adding adapter check
* linting
* simplifying check
* linting v2
* config fix -___-
2025-03-21 10:18:28 -04:00
NanoCode012
f8de8bb4f2
chore(doc): add instructions on adding custom integrations ( #2422 ) [skip ci]
...
* chore(doc): add instructions on adding custom integrations
* chore: add warning help
* feat: add note about integration path
* fix: adjust text per suggestion
2025-03-21 10:18:01 -04:00
hugo
8e604848a4
add run on novita ai ( #2421 ) [skip ci]
...
* add run on novita ai
* Revert "add run on novita ai"
This reverts commit 4d5df1ac6b .
* add run axolotl on novita ai
2025-03-21 10:17:47 -04:00
Wing Lian
aae4337f40
add 12.8.1 cuda to the base matrix ( #2426 )
...
* add 12.8.1 cuda to the base matrix
* use nightly
* bump deepspeed and set no binary
* deepspeed binary fixes hopefully
* install deepspeed by itself
* multiline fix
* make sure ninja is installed
* try with reversion of packaging/setuptools/wheel install
* use license instead of license-file
* try rolling back packaging and setuptools versions
* comment out license for validation for now
* make sure packaging version is consistent
* more parity across tests and docker images for packaging/setuptools
2025-03-21 10:17:25 -04:00
Wing Lian
38df5a36ea
bump HF versions except for trl ( #2427 )
2025-03-20 10:22:05 -04:00
Wing Lian
4d92a68a96
use default torch fused adamw optimizer as default as adamw_hf is deprecated ( #2425 )
...
* use default torch fused adamw optimizer as default as adamw_hf is deprecated
* make sure to have latest packaging installed
* bump packagingin requirements.txt too
2025-03-19 23:58:33 -04:00
SicariusSicariiStuff
85147ec430
Update README.md ( #2360 )
...
* Update README.md
wheel is needed
* feat: add ninja, setuptools, packing to installation steps
* fix: add missing instruction
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
2025-03-17 08:39:17 -04:00
NanoCode012
51cd409488
Feat: minor docs improvements for RLHF and faq on embeddings ( #2401 ) [skip ci]
...
* feat: add doc on shrink_embeddings and custom calling
* chore: rename inference doc
* fix: clarify same config is used for all cli
* chore: rearrange order inference qmd
* feat: add simpo to doc
* fix: update defaults
* feat: add rl configs to doc
* fix: ensure beta consistent with trl.beta
* fix: clarify about lora/fft
* chore: rename title
* chore: fix language
* feat: move config reference higher
* Update docs/getting-started.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
* Update docs/rlhf.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-17 08:39:04 -04:00
NanoCode012
7235123d44
chore(docs): add cookbook/blog link to docs ( #2410 ) [skip ci]
2025-03-17 08:38:19 -04:00
Wing Lian
4f5eb42a73
remove reference to deprecated import ( #2407 )
2025-03-15 08:49:41 -04:00
Wing Lian
fbe54be6b8
only validate hf user token on rank 0 ( #2408 )
2025-03-13 23:29:06 -04:00
Wing Lian
04f6324833
build cloud images with torch 2.6.0 ( #2413 )
...
* build cloud images with torch 2.6.0
* nightlies too
2025-03-13 23:28:51 -04:00
Wing Lian
f0072f3b9d
use max of 32 dataset processes if not explicit ( #2403 )
...
* use max of 32 dataset processes if not explicit
* change alternate min val for consistency
2025-03-11 12:02:58 -04:00
Wing Lian
59899b9817
pass additional info for fix untrained tokens when using distributed + offloading ( #2388 )
...
* pass additional info for fix untrained tokens when using distributed + offloading
* use latest version of vendored lib
* use v0.0.5 of contribs lgpl
* fix for no bad tokens and add tests
* use release
* add multigpu test too
* make sure the multigpu zero3 test actually uses zero3
2025-03-11 12:02:43 -04:00
NanoCode012
4a736986fa
fix(modal): add git pull when getting branch files ( #2399 )
2025-03-10 15:14:41 -04:00
Wing Lian
5d0f110a3b
include iproute2 and nvtop in cloud image ( #2393 )
2025-03-10 15:13:38 -04:00
NanoCode012
83f8698b8a
fix: create mount folder on modal if not exist ( #2390 )
2025-03-10 16:27:42 +07:00
xzuyn
60a11a6410
Use Latest Cut Cross Entropy ( #2392 )
...
* Update __init__.py
* Update README.md
* Update cutcrossentropy_install.py
* add test
2025-03-10 16:26:40 +07:00
NanoCode012
46a045e528
chore(doc): add faq when having no default chat_template ( #2398 )
...
* chore(doc): add faq when having no default chat_template
* Update docs/dataset-formats/conversation.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
* Update docs/faq.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-10 16:25:50 +07:00
NanoCode012
3b477e08a0
feat(doc): add more info on RewardModel datasets ( #2391 )
...
* fix: reduce title size
* feat(doc): add rm dataset info
* Update docs/reward_modelling.qmd following suggestion
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-10 16:25:31 +07:00
NanoCode012
16dc6ee68d
refactor: trl grpo configs to have descriptions ( #2386 )
...
* refactor: trl grpo configs to have descriptions
* chore: caps
2025-03-07 08:58:53 -05:00
Wing Lian
fa7c79b3b9
remove lion-pytorch as it's already handled upstream ( #2389 )
2025-03-07 08:58:15 -05:00
Wing Lian
ae66374156
Optimizer refactor and add Muon support ( #2367 )
...
* add muon optimizer
optimizer_cls_and_kwargs is on trainer_kwargs
only add adamw_kwargs if they're non-null
fix mocks
better handling of override and check the optimizer
unwrap optimizer
* fix import
2025-03-06 11:49:19 -05:00
Wing Lian
5e21b1a9da
various fixes 20250305 ( #2384 )
...
* various validation fixes
* fix check for non-truthy value
2025-03-06 11:48:44 -05:00
mhenrichsen
575e5f28ec
Update Tokenizer Overrides Handling in models.py ( #1549 )
...
* override special tokens mock code
* fix(doc): remove duplicate config
* feat: replace added_tokens in tokenizer and add test
* make sure to run tokenizer modification on rank 0 only
* use is local main process instead
* feat: rename config
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-03-05 11:15:12 -05:00
xzuyn
0134093acc
Add REX LR Scheduler ( #2380 )
...
* Update trainer_builder.py
* Update base.py
* Update __init__.py
* Update base.py
* Update base.py
* Update config.qmd
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* lint
* lint
* lint
* lint
* lint
* lint
* Update base.py
* Update base.py
* lint
* Update base.py
* Update base.py
* Move RexLR to `schedulers.py`
* Remove RexLR from `base.py`
* Fix tooltip formatting
* lint
* Create test_schedulers.py
* Use a default optimizer in test
* lint
* lint
* Add `warmup_steps` and `cosine_min_lr_ratio` to test
* lint
2025-03-05 10:26:11 -05:00
NanoCode012
d4de93a7bb
feat(grpo): add reward_weights config and refactor ( #2365 )
2025-03-05 10:02:08 -05:00
NanoCode012
c8191394e9
fix(doc): add missing low_cpu_mem_usage config to docs ( #2369 ) [skip ci]
2025-03-05 10:01:44 -05:00
NanoCode012
f18231c653
chore(doc): add clarification about mpi4py error on single gpu deepspeed ( #2383 ) [skip ci]
...
* chore(doc): add clarification about mpi4py error on single gpu deepspeed
* fix: lint
2025-03-05 10:01:28 -05:00
NanoCode012
9ed4f6b3aa
feat(doc): document drop_system_message and clarify limitation ( #2381 ) [skip ci]
2025-03-05 10:01:16 -05:00
NanoCode012
05dddfc41d
feat(doc): add docker images explanation ( #2379 ) [skip ci]
...
* feat(doc): add docker images explanation
* chore: add link to dockerhub
2025-03-05 10:01:00 -05:00
NanoCode012
8e30917440
chore(docs): remove phorm ( #2378 ) [skip ci]
2025-03-05 10:00:50 -05:00
NanoCode012
d883b11b6f
fix(doc): add installation for cce to docs ( #2375 ) [skip ci]
...
* fix(doc): add installation for cce to docs
* fix: format
2025-03-05 10:00:39 -05:00
Dan Saunders
f4910dd2ea
train.py refactor (#2371 )
...
* refactor train.py
* updates
* update
* combine like functions
* review comments
2025-03-05 08:58:33 -05:00
NanoCode012
75cbd15301
Fix(doc): address missing doc changes ( #2362 )
...
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl (<nil>, 124, 12.4.1, 3.11, 2.6.0) (push) Has been cancelled
ci-cd / build-axolotl (vllm, 124, 12.4.1, true, 3.11, 2.5.1) (push) Has been cancelled
publish pypi / Create Release (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud (<nil>, 124, 12.4.1, true, 3.11, 2.5.1) (push) Has been cancelled
ci-cd / build-axolotl-cloud-no-tmux (<nil>, 124, 12.4.1, 3.11, 2.4.1) (push) Has been cancelled
publish pypi / Upload release to PyPI (push) Has been cancelled
* fix: add multiple tips about eos_token masking
* fix: format dataset preprocessing doc
* Update docs/dataset-formats/conversation.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
v0.7.1
2025-02-25 13:50:02 -05:00
NanoCode012
2efe1b4c09
Feat(doc): Reorganize documentation, fix broken syntax, update notes ( #2348 )
...
* feat(doc): organize docs, add to menu bar, fix broken formatting
* feat: add link to custom integrations
* feat: update readme for integrations to include citations and repo link
* chore: update lm_eval info
* chore: use fullname
* Update docs/cli.qmd per suggestion
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
* feat: add sweep doc
* feat: add kd doc
* fix: remove toc
* fix: update deprecation
* feat: add more info about chat_template issues
* fix: heading level
* fix: shell->bash code block
* fix: ray link
* fix(doc): heading level, header links, formatting
* feat: add grpo docs
* feat: add style changes
* fix: wrong cli arg for lm-eval
* fix: remove old run method
* feat: load custom integration doc dynamically
* fix: remove old cli way
* fix: toc
* fix: minor formatting
---------
Co-authored-by: Dan Saunders <danjsaund@gmail.com >
2025-02-25 16:09:37 +07:00
NanoCode012
1110a37e21
feat: add deepseek_v3 sample packing ( #2230 )
2025-02-24 15:03:15 -05:00
Wing Lian
9850f42204
bump liger to 0.5.3 ( #2353 )
2025-02-24 12:40:54 -05:00
Matt Baker
00fc8109e4
Correctly reference mount paths ( #2347 )
...
* Correctly reference mount paths
* Also fix mount paths in lm_eval
* chore: lint
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-02-24 11:12:57 -05:00
Wing Lian
2d5826f544
Relicense the logprob KD loss functions as Apache 2.0 ( #2358 )
2025-02-23 12:31:35 -05:00
Wing Lian
a4170030ab
don't install extraneous old version of pydantic in ci and make sre to run multigpu ci ( #2355 )
2025-02-21 22:06:29 -05:00