Dan Saunders
a26985c53c
small changes
2025-03-21 16:36:17 +00:00
Dan Saunders
c1a58339e8
add SP doc, review comments
2025-03-21 16:36:17 +00:00
Dan Saunders
411df76a97
bugfix
2025-03-21 16:36:17 +00:00
Dan Saunders
a09d1ccbf2
removing print statement
2025-03-21 16:36:17 +00:00
Dan Saunders
2727d86544
non-seq2se1 collator fix
2025-03-21 16:36:17 +00:00
Dan Saunders
64c203cdef
sampler / dataloader refactor
2025-03-21 16:36:17 +00:00
Dan Saunders
7d7042f602
test fix
2025-03-21 16:36:17 +00:00
Dan Saunders
d187f1f8e2
using field validator instead of model validator
2025-03-21 16:36:17 +00:00
Dan Saunders
e1a02a32b5
fix
2025-03-21 16:36:17 +00:00
Dan Saunders
a6ef6c7764
fix
2025-03-21 16:36:17 +00:00
Dan Saunders
cb3a9e99a3
gracefully handle no ring-flash-attn
2025-03-21 16:36:17 +00:00
Dan Saunders
03027cf6bf
pernicious Fire CLI bugfix
2025-03-21 16:36:16 +00:00
Dan Saunders
0ade60d455
another import scoping change
2025-03-21 16:35:56 +00:00
Dan Saunders
02e1a42f04
scoping down problematic import
2025-03-21 16:35:56 +00:00
Dan Saunders
919b88f11b
update config.qmd and rename option
2025-03-21 16:35:55 +00:00
Dan Saunders
345a9dd831
removing some obvious comments
2025-03-21 16:35:38 +00:00
Dan Saunders
4ff97bc9d4
eval dataloader and sampler changes
2025-03-21 16:35:38 +00:00
Dan Saunders
d0e178d52f
remove debug logs and simplify
2025-03-21 16:35:38 +00:00
Dan Saunders
5731cdc0cf
fixing sample packing
2025-03-21 16:35:38 +00:00
Dan Saunders
b7738d57c4
working multi-group SP
2025-03-21 16:35:38 +00:00
Dan Saunders
698e599bf7
precommit fixes
2025-03-21 16:35:38 +00:00
Dan Saunders
1d339e4007
fixes
2025-03-21 16:35:38 +00:00
Dan Saunders
4190ad0647
updates
2025-03-21 16:35:36 +00:00
Dan Saunders
51c326150b
pytest
2025-03-21 16:35:10 +00:00
Dan Saunders
14baaf6e0a
updates
2025-03-21 16:35:10 +00:00
Dan Saunders
f487910444
removing unused code
2025-03-21 16:35:08 +00:00
Dan Saunders
7efc787ac8
cleanup
2025-03-21 16:34:12 +00:00
Dan Saunders
dce61cdab1
progress on ring attn impl
2025-03-21 16:34:12 +00:00
Dan Saunders
bd952de9d2
progress on ring attn impl
2025-03-21 16:34:10 +00:00
Dan Saunders
3f8a43cab6
adding easy_context as integration for now
2025-03-21 16:33:46 +00:00
Dan Saunders
113e9cd193
Autodoc generation with quartodoc ( #2419 )
...
* quartodoc integration
* quartodoc progress
* deletions
* Update docs/.gitignore to exclude auto-generated API documentation files
* Fix
* more autodoc progress
* moving reference up near the top of the sidebar
* fix broken link
* update to reflect recent changes
* pydantic models refactor + add to autodoc + fixes
* fix
* shrinking header sizes
* fix accidental change
* include quartodoc build step
* update pre-commit version
* update pylint
* pre-commit
---------
Co-authored-by: Dan Saunders <dan@axolotl.ai >
2025-03-21 12:26:47 -04:00
Dan Saunders
c907ac173e
adding pre-commit auto-update GH action and bumping plugin versions ( #2428 )
...
* adding pre-commit auto-update GH action and bumping plugin versions
* running updated pre-commit plugins
* sorry to revert, but pylint complained
* Update .pre-commit-config.yaml
Co-authored-by: Wing Lian <wing.lian@gmail.com >
---------
Co-authored-by: Dan Saunders <dan@axolotl.ai >
Co-authored-by: Wing Lian <wing.lian@gmail.com >
2025-03-21 11:02:43 -04:00
salman
187227d837
Fixing KTO+QLoRA+multi-GPU ( #2420 )
...
* WIP
* removing artifacts
* adding error
* adding adapter check
* linting
* simplifying check
* linting v2
* config fix -___-
2025-03-21 10:18:28 -04:00
Wing Lian
38df5a36ea
bump HF versions except for trl ( #2427 )
2025-03-20 10:22:05 -04:00
Wing Lian
4d92a68a96
use default torch fused adamw optimizer as default as adamw_hf is deprecated ( #2425 )
...
* use default torch fused adamw optimizer as default as adamw_hf is deprecated
* make sure to have latest packaging installed
* bump packagingin requirements.txt too
2025-03-19 23:58:33 -04:00
NanoCode012
51cd409488
Feat: minor docs improvements for RLHF and faq on embeddings ( #2401 ) [skip ci]
...
* feat: add doc on shrink_embeddings and custom calling
* chore: rename inference doc
* fix: clarify same config is used for all cli
* chore: rearrange order inference qmd
* feat: add simpo to doc
* fix: update defaults
* feat: add rl configs to doc
* fix: ensure beta consistent with trl.beta
* fix: clarify about lora/fft
* chore: rename title
* chore: fix language
* feat: move config reference higher
* Update docs/getting-started.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
* Update docs/rlhf.qmd
Co-authored-by: salman <salman.mohammadi@outlook.com >
---------
Co-authored-by: salman <salman.mohammadi@outlook.com >
2025-03-17 08:39:04 -04:00
Wing Lian
4f5eb42a73
remove reference to deprecated import ( #2407 )
2025-03-15 08:49:41 -04:00
Wing Lian
fbe54be6b8
only validate hf user token on rank 0 ( #2408 )
2025-03-13 23:29:06 -04:00
Wing Lian
f0072f3b9d
use max of 32 dataset processes if not explicit ( #2403 )
...
* use max of 32 dataset processes if not explicit
* change alternate min val for consistency
2025-03-11 12:02:58 -04:00
Wing Lian
59899b9817
pass additional info for fix untrained tokens when using distributed + offloading ( #2388 )
...
* pass additional info for fix untrained tokens when using distributed + offloading
* use latest version of vendored lib
* use v0.0.5 of contribs lgpl
* fix for no bad tokens and add tests
* use release
* add multigpu test too
* make sure the multigpu zero3 test actually uses zero3
2025-03-11 12:02:43 -04:00
NanoCode012
4a736986fa
fix(modal): add git pull when getting branch files ( #2399 )
2025-03-10 15:14:41 -04:00
NanoCode012
83f8698b8a
fix: create mount folder on modal if not exist ( #2390 )
2025-03-10 16:27:42 +07:00
xzuyn
60a11a6410
Use Latest Cut Cross Entropy ( #2392 )
...
* Update __init__.py
* Update README.md
* Update cutcrossentropy_install.py
* add test
2025-03-10 16:26:40 +07:00
NanoCode012
16dc6ee68d
refactor: trl grpo configs to have descriptions ( #2386 )
...
* refactor: trl grpo configs to have descriptions
* chore: caps
2025-03-07 08:58:53 -05:00
Wing Lian
fa7c79b3b9
remove lion-pytorch as it's already handled upstream ( #2389 )
2025-03-07 08:58:15 -05:00
Wing Lian
ae66374156
Optimizer refactor and add Muon support ( #2367 )
...
* add muon optimizer
optimizer_cls_and_kwargs is on trainer_kwargs
only add adamw_kwargs if they're non-null
fix mocks
better handling of override and check the optimizer
unwrap optimizer
* fix import
2025-03-06 11:49:19 -05:00
Wing Lian
5e21b1a9da
various fixes 20250305 ( #2384 )
...
* various validation fixes
* fix check for non-truthy value
2025-03-06 11:48:44 -05:00
mhenrichsen
575e5f28ec
Update Tokenizer Overrides Handling in models.py ( #1549 )
...
* override special tokens mock code
* fix(doc): remove duplicate config
* feat: replace added_tokens in tokenizer and add test
* make sure to run tokenizer modification on rank 0 only
* use is local main process instead
* feat: rename config
---------
Co-authored-by: NanoCode012 <nano@axolotl.ai >
Co-authored-by: Wing Lian <wing@axolotl.ai >
2025-03-05 11:15:12 -05:00
xzuyn
0134093acc
Add REX LR Scheduler ( #2380 )
...
* Update trainer_builder.py
* Update base.py
* Update __init__.py
* Update base.py
* Update base.py
* Update config.qmd
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* Update base.py
* lint
* lint
* lint
* lint
* lint
* lint
* Update base.py
* Update base.py
* lint
* Update base.py
* Update base.py
* Move RexLR to `schedulers.py`
* Remove RexLR from `base.py`
* Fix tooltip formatting
* lint
* Create test_schedulers.py
* Use a default optimizer in test
* lint
* lint
* Add `warmup_steps` and `cosine_min_lr_ratio` to test
* lint
2025-03-05 10:26:11 -05:00
NanoCode012
d4de93a7bb
feat(grpo): add reward_weights config and refactor ( #2365 )
2025-03-05 10:02:08 -05:00