Commit Graph

1028 Commits

Author SHA1 Message Date
Dan Saunders
a26985c53c small changes 2025-03-21 16:36:17 +00:00
Dan Saunders
c1a58339e8 add SP doc, review comments 2025-03-21 16:36:17 +00:00
Dan Saunders
411df76a97 bugfix 2025-03-21 16:36:17 +00:00
Dan Saunders
a09d1ccbf2 removing print statement 2025-03-21 16:36:17 +00:00
Dan Saunders
2727d86544 non-seq2se1 collator fix 2025-03-21 16:36:17 +00:00
Dan Saunders
64c203cdef sampler / dataloader refactor 2025-03-21 16:36:17 +00:00
Dan Saunders
7d7042f602 test fix 2025-03-21 16:36:17 +00:00
Dan Saunders
d187f1f8e2 using field validator instead of model validator 2025-03-21 16:36:17 +00:00
Dan Saunders
e1a02a32b5 fix 2025-03-21 16:36:17 +00:00
Dan Saunders
a6ef6c7764 fix 2025-03-21 16:36:17 +00:00
Dan Saunders
cb3a9e99a3 gracefully handle no ring-flash-attn 2025-03-21 16:36:17 +00:00
Dan Saunders
03027cf6bf pernicious Fire CLI bugfix 2025-03-21 16:36:16 +00:00
Dan Saunders
0ade60d455 another import scoping change 2025-03-21 16:35:56 +00:00
Dan Saunders
02e1a42f04 scoping down problematic import 2025-03-21 16:35:56 +00:00
Dan Saunders
919b88f11b update config.qmd and rename option 2025-03-21 16:35:55 +00:00
Dan Saunders
345a9dd831 removing some obvious comments 2025-03-21 16:35:38 +00:00
Dan Saunders
4ff97bc9d4 eval dataloader and sampler changes 2025-03-21 16:35:38 +00:00
Dan Saunders
d0e178d52f remove debug logs and simplify 2025-03-21 16:35:38 +00:00
Dan Saunders
5731cdc0cf fixing sample packing 2025-03-21 16:35:38 +00:00
Dan Saunders
b7738d57c4 working multi-group SP 2025-03-21 16:35:38 +00:00
Dan Saunders
698e599bf7 precommit fixes 2025-03-21 16:35:38 +00:00
Dan Saunders
1d339e4007 fixes 2025-03-21 16:35:38 +00:00
Dan Saunders
4190ad0647 updates 2025-03-21 16:35:36 +00:00
Dan Saunders
51c326150b pytest 2025-03-21 16:35:10 +00:00
Dan Saunders
14baaf6e0a updates 2025-03-21 16:35:10 +00:00
Dan Saunders
f487910444 removing unused code 2025-03-21 16:35:08 +00:00
Dan Saunders
7efc787ac8 cleanup 2025-03-21 16:34:12 +00:00
Dan Saunders
dce61cdab1 progress on ring attn impl 2025-03-21 16:34:12 +00:00
Dan Saunders
bd952de9d2 progress on ring attn impl 2025-03-21 16:34:10 +00:00
Dan Saunders
3f8a43cab6 adding easy_context as integration for now 2025-03-21 16:33:46 +00:00
Dan Saunders
113e9cd193 Autodoc generation with quartodoc (#2419)
* quartodoc integration

* quartodoc progress

* deletions

* Update docs/.gitignore to exclude auto-generated API documentation files

* Fix

* more autodoc progress

* moving reference up near the top of the sidebar

* fix broken link

* update to reflect recent changes

* pydantic models refactor + add to autodoc + fixes

* fix

* shrinking header sizes

* fix accidental change

* include quartodoc build step

* update pre-commit version

* update pylint

* pre-commit

---------

Co-authored-by: Dan Saunders <dan@axolotl.ai>
2025-03-21 12:26:47 -04:00
Dan Saunders
c907ac173e adding pre-commit auto-update GH action and bumping plugin versions (#2428)
* adding pre-commit auto-update GH action and bumping plugin versions

* running updated pre-commit plugins

* sorry to revert, but pylint complained

* Update .pre-commit-config.yaml

Co-authored-by: Wing Lian <wing.lian@gmail.com>

---------

Co-authored-by: Dan Saunders <dan@axolotl.ai>
Co-authored-by: Wing Lian <wing.lian@gmail.com>
2025-03-21 11:02:43 -04:00
salman
187227d837 Fixing KTO+QLoRA+multi-GPU (#2420)
* WIP

* removing artifacts

* adding error

* adding adapter check

* linting

* simplifying check

* linting v2

* config fix -___-
2025-03-21 10:18:28 -04:00
Wing Lian
38df5a36ea bump HF versions except for trl (#2427) 2025-03-20 10:22:05 -04:00
Wing Lian
4d92a68a96 use default torch fused adamw optimizer as default as adamw_hf is deprecated (#2425)
* use default torch fused adamw optimizer as default as adamw_hf is deprecated

* make sure to have latest packaging installed

* bump packagingin requirements.txt too
2025-03-19 23:58:33 -04:00
NanoCode012
51cd409488 Feat: minor docs improvements for RLHF and faq on embeddings (#2401) [skip ci]
* feat: add doc on shrink_embeddings and custom calling

* chore: rename inference doc

* fix: clarify same config is used for all cli

* chore: rearrange order inference qmd

* feat: add simpo to doc

* fix: update defaults

* feat: add rl configs to doc

* fix: ensure beta consistent with trl.beta

* fix: clarify about lora/fft

* chore: rename title

* chore: fix language

* feat: move config reference higher

* Update docs/getting-started.qmd

Co-authored-by: salman <salman.mohammadi@outlook.com>

* Update docs/rlhf.qmd

Co-authored-by: salman <salman.mohammadi@outlook.com>

---------

Co-authored-by: salman <salman.mohammadi@outlook.com>
2025-03-17 08:39:04 -04:00
Wing Lian
4f5eb42a73 remove reference to deprecated import (#2407) 2025-03-15 08:49:41 -04:00
Wing Lian
fbe54be6b8 only validate hf user token on rank 0 (#2408) 2025-03-13 23:29:06 -04:00
Wing Lian
f0072f3b9d use max of 32 dataset processes if not explicit (#2403)
* use max of 32 dataset processes if not explicit

* change alternate min val for consistency
2025-03-11 12:02:58 -04:00
Wing Lian
59899b9817 pass additional info for fix untrained tokens when using distributed + offloading (#2388)
* pass additional info for fix untrained tokens when using distributed + offloading

* use latest version of vendored lib

* use v0.0.5 of contribs lgpl

* fix for no bad tokens and add tests

* use release

* add multigpu test too

* make sure the multigpu zero3 test actually uses zero3
2025-03-11 12:02:43 -04:00
NanoCode012
4a736986fa fix(modal): add git pull when getting branch files (#2399) 2025-03-10 15:14:41 -04:00
NanoCode012
83f8698b8a fix: create mount folder on modal if not exist (#2390) 2025-03-10 16:27:42 +07:00
xzuyn
60a11a6410 Use Latest Cut Cross Entropy (#2392)
* Update __init__.py

* Update README.md

* Update cutcrossentropy_install.py

* add test
2025-03-10 16:26:40 +07:00
NanoCode012
16dc6ee68d refactor: trl grpo configs to have descriptions (#2386)
* refactor: trl grpo configs to have descriptions

* chore: caps
2025-03-07 08:58:53 -05:00
Wing Lian
fa7c79b3b9 remove lion-pytorch as it's already handled upstream (#2389) 2025-03-07 08:58:15 -05:00
Wing Lian
ae66374156 Optimizer refactor and add Muon support (#2367)
* add muon optimizer

optimizer_cls_and_kwargs is on trainer_kwargs
only add adamw_kwargs if they're non-null
fix mocks
better handling of override and check the optimizer
unwrap optimizer

* fix import
2025-03-06 11:49:19 -05:00
Wing Lian
5e21b1a9da various fixes 20250305 (#2384)
* various validation fixes

* fix check for non-truthy value
2025-03-06 11:48:44 -05:00
mhenrichsen
575e5f28ec Update Tokenizer Overrides Handling in models.py (#1549)
* override special tokens mock code

* fix(doc): remove duplicate config

* feat: replace added_tokens in tokenizer and add test

* make sure to run tokenizer modification on rank 0 only

* use is local main process instead

* feat: rename config

---------

Co-authored-by: NanoCode012 <nano@axolotl.ai>
Co-authored-by: Wing Lian <wing@axolotl.ai>
2025-03-05 11:15:12 -05:00
xzuyn
0134093acc Add REX LR Scheduler (#2380)
* Update trainer_builder.py

* Update base.py

* Update __init__.py

* Update base.py

* Update base.py

* Update config.qmd

* Update base.py

* Update base.py

* Update base.py

* Update base.py

* Update base.py

* Update base.py

* Update base.py

* lint

* lint

* lint

* lint

* lint

* lint

* Update base.py

* Update base.py

* lint

* Update base.py

* Update base.py

* Move RexLR to `schedulers.py`

* Remove RexLR from `base.py`

* Fix tooltip formatting

* lint

* Create test_schedulers.py

* Use a default optimizer in test

* lint

* lint

* Add `warmup_steps` and `cosine_min_lr_ratio` to test

* lint
2025-03-05 10:26:11 -05:00
NanoCode012
d4de93a7bb feat(grpo): add reward_weights config and refactor (#2365) 2025-03-05 10:02:08 -05:00