Wing Lian
ab690f3f01
improve check for batched
2025-01-14 22:47:46 -05:00
Wing Lian
47932f21c4
fix reward trainer calls for tokenization
2025-01-14 22:47:46 -05:00
Wing Lian
808328e041
reward can use same batch check
2025-01-14 22:47:46 -05:00
Wing Lian
6784822cfb
tweak check for batched prompt data
2025-01-14 22:47:46 -05:00
Wing Lian
684b38291f
ensure that batch vs single is done properly
2025-01-14 22:47:46 -05:00
Wing Lian
01896b1bde
improve iterable support
2025-01-14 22:47:46 -05:00
Wing Lian
e659c01646
support streaming for processing sft datasts?
2025-01-14 22:47:45 -05:00
Wing Lian
204d6c43b4
make loss torch script compat
2025-01-14 22:47:45 -05:00
Wing Lian
d3c2b7ce9d
kd sample packing
2025-01-14 22:47:45 -05:00
Wing Lian
93dfff92f1
be a bit pickier about loading dynamic prompt strategies
2025-01-14 22:47:45 -05:00
Wing Lian
6e409d2d88
more info on preprocess for kd and fix import
2025-01-14 22:47:45 -05:00
Wing Lian
d5bc214300
remove duplicate code
2025-01-14 22:47:45 -05:00
Wing Lian
92c6c1087e
add copyrights
2025-01-14 22:47:45 -05:00
Wing Lian
feed96f95e
increase logging around loading plugins
2025-01-14 22:47:44 -05:00
Wing Lian
cba6165ae1
make plugin setup concise
2025-01-14 22:47:44 -05:00
Wing Lian
cdfcd69afa
remove moved class from import
2025-01-14 22:47:44 -05:00
Wing Lian
885653d52e
move more things to kd plugin
2025-01-14 22:47:44 -05:00
Wing Lian
27faacbf5a
refactor kd chat template loader
2025-01-14 22:47:44 -05:00
Wing Lian
c51b0337c1
support for custom trainer classes from plugins
2025-01-14 22:47:44 -05:00
Wing Lian
fa055f9f69
handle token/logprob shifting
2025-01-14 22:47:43 -05:00
Wing Lian
f60c623af0
remove references to triton kd for now
2025-01-14 22:47:43 -05:00
Wing Lian
746891eb5c
add license block
2025-01-14 22:47:43 -05:00
Wing Lian
f09b5da60b
refactor so we can easily add new loss functions
2025-01-14 22:47:43 -05:00
Wing Lian
689e1c10ba
chore: lint
2025-01-14 22:47:43 -05:00
Wing Lian
a5c085e003
var naming and add todo
2025-01-14 22:47:43 -05:00
Wing Lian
63146300b7
fix kd loss so it's causal (fixes repeating tokens)
2025-01-14 22:47:43 -05:00
Wing Lian
ca5e397fc5
use kd_alpha in the correct loss method
2025-01-14 22:47:42 -05:00
Wing Lian
3416302b0d
hash for temperature too
2025-01-14 22:47:42 -05:00
Wing Lian
7366efc4ca
better rescaling for temperatures
2025-01-14 22:47:42 -05:00
Wing Lian
d8d817eaed
don't use triton for now
2025-01-14 22:47:42 -05:00
Wing Lian
c0757e8a20
fix kwarg
2025-01-14 22:47:42 -05:00
Wing Lian
e565694914
v3
2025-01-14 22:47:42 -05:00
Wing Lian
081928e55b
no torch.tensor
2025-01-14 22:47:42 -05:00
Wing Lian
dc90c93894
no log etc
2025-01-14 22:47:41 -05:00
Wing Lian
18a46c338a
no torch.exp inside triton kernel
2025-01-14 22:47:41 -05:00
Wing Lian
119d586cf4
v2 trial
2025-01-14 22:47:41 -05:00
Wing Lian
c73acd7de0
no where support
2025-01-14 22:47:41 -05:00
Wing Lian
0b59a242d4
triton wip
2025-01-14 22:47:41 -05:00
Wing Lian
ed490517da
chore: lint
2025-01-14 22:47:41 -05:00
Wing Lian
00ce77e7ef
make sure to multiply against the correct loss
2025-01-14 22:47:41 -05:00
Wing Lian
ae545e0165
cross entropy loss coefficient during KD
2025-01-14 22:47:40 -05:00
Wing Lian
b592c05b93
flipped the slice
2025-01-14 22:47:40 -05:00
Wing Lian
7fe0ad088b
make it work
2025-01-14 22:47:40 -05:00
Wing Lian
ddcf5c68b3
handle padding/collation for KD datasets
2025-01-14 22:47:40 -05:00
Wing Lian
e633a12dbe
make batch smaller
2025-01-14 22:47:40 -05:00
Wing Lian
d584354ee4
filter bad rows
2025-01-14 22:47:40 -05:00
Wing Lian
303cfa71aa
KD dataset loading and KD with logprobs
2025-01-14 22:47:40 -05:00
Wing Lian
88b3198894
refactor trainer to prevent circular dependencies later
...
fix loader default
2025-01-14 22:47:39 -05:00
jwongTensora
8606093921
fix for indexing error from token/embeddings mismatch ( #2257 )
...
Co-authored-by: jwong <jwongTensora@gmail.com >
2025-01-14 22:09:29 -05:00
NanoCode012
cba5a457d9
fix: use text_column even when not packing for pretraining ( #2254 )
...
* fix: use text_column even when not packing for pretraining
* feat: update test to check when not packing
* chore: lint
* Update src/axolotl/utils/data/pretraining.py
Co-authored-by: Wing Lian <wing.lian@gmail.com >
---------
Co-authored-by: Wing Lian <wing@axolotl.ai >
Co-authored-by: Wing Lian <wing.lian@gmail.com >
2025-01-14 22:08:56 -05:00