Wing Lian
|
b448c77148
|
address pr feedback
|
2023-08-29 22:45:22 -07:00 |
|
Wing Lian
|
c820d04669
|
gptq doesn't play well with sample packing
|
2023-08-29 12:15:31 -07:00 |
|
Wing Lian
|
a29560004b
|
more tweaks and add yml
|
2023-08-29 12:02:00 -07:00 |
|
Birch-san
|
8e197f6fb4
|
pad_to_worst_case_seq_len boolean, for testing memory limits (#498)
* pad_to_worst_case_seq_len boolean, for testing memory limits
* remove collator_pad_to_longest option since it does nothing
see docs: https://huggingface.co/docs/transformers/main_classes/data_collator#transformers.DataCollatorWithPadding.padding
True and "longest" mean the same thing
* rename to `pad_to_sequence_len, and ensure 64 alignment
---------
Co-authored-by: Aman Karmani <aman@tmm1.net>
|
2023-08-28 18:47:16 -04:00 |
|
mhenrichsen
|
35130711d6
|
Feat(cfg): Add code-llama configs for all sizes (#479)
* configs for all sizes
* update tokenizer type
---------
Co-authored-by: mhenrichsen <some_email@hey.com>
|
2023-08-27 10:20:17 +09:00 |
|
Charles O. Goddard
|
fe4d6baf92
|
Add example Llama 2 ReLoRA config (#471)
* Add example Llama 2 ReLoRA config
* Use adamw_bnb_8bit in example relora config
|
2023-08-27 10:08:34 +09:00 |
|
Wing Lian
|
cb9797ef5a
|
improve llama pad token handling (#475)
* improve llama pad token handling
* tweak logic to not clobber
|
2023-08-24 13:20:35 -04:00 |
|
Wing Lian
|
1687be6a35
|
don't use mask expansion for inference (#392)
|
2023-08-14 20:52:54 -04:00 |
|
mhenrichsen
|
fdffef5940
|
new llama-2 default settings (#370)
* new default settings
* fix whitespace
* rm max packed sequence length
---------
Co-authored-by: Mads Henrichsen <mads@BrbartiendeMads.lan>
|
2023-08-14 17:39:09 +09:00 |
|
Morgan McGuire
|
7019509daa
|
Add wandb_entity to wandb options, update example configs, update README (#361)
* Update wandb_entity and add wandb descriptions
* add wandb to config section
* remove trailing whitespace for pre-commit hook
* remove trailing whitespace for pre-commit hook
---------
Co-authored-by: Morgan McGuire <morganmcguire@Morgans-MacBook-Pro.local>
Co-authored-by: Wing Lian <wing.lian@gmail.com>
|
2023-08-12 12:17:11 -04:00 |
|
Aman Karmani
|
36fefcf94b
|
set group_by_length to false in examples
|
2023-08-06 23:59:09 -07:00 |
|
mhenrichsen
|
dc71d8872a
|
feat/llama-2 examples (#319)
* qlora llama-2
* qlora llama-2
* linting
* readme
* lora added
* linting
* change group_by_length
* 13b fitting on 24gb
* grouped lengths true
* add pad token
* change out dir
---------
Co-authored-by: Mads Henrichsen <mads@Brbar-tilhrende-Mads.local>
|
2023-08-03 19:22:48 +09:00 |
|
Ethan Smith
|
38811434e6
|
Add XGen info to README and example config
|
2023-07-21 00:44:50 -07:00 |
|
Steffen Röcker
|
945c4191a3
|
Use AutoTokenizer for redpajama example
|
2023-06-14 20:09:26 +02:00 |
|
Wing Lian
|
16bb6276a5
|
Merge pull request #92 from OpenAccess-AI-Collective/flash-optimum
add support for opimum bettertransformers
|
2023-06-14 07:50:15 -04:00 |
|
Wing Lian
|
fd2c9814c9
|
Merge branch 'main' into flash-optimum
|
2023-06-12 13:12:15 -04:00 |
|
Wing Lian
|
2ba4ae8f46
|
tweak config to work
|
2023-06-12 10:07:18 -04:00 |
|
Wing Lian
|
94f310c7a6
|
Merge pull request #193 from OpenAccess-AI-Collective/config-fixes-20230612
config fixes
|
2023-06-12 08:24:52 -04:00 |
|
NanoCode012
|
52cde69288
|
Fix config path after config moved
|
2023-06-12 17:06:15 +09:00 |
|
Wing Lian
|
9a58e99e81
|
config fixes
|
2023-06-12 01:52:58 -04:00 |
|
Wing Lian
|
6b3f509d9e
|
forgot to add this file
|
2023-06-11 11:50:12 -04:00 |
|
Wing Lian
|
d0d7eaa4f3
|
update openllama and clean up paths
|
2023-06-11 11:03:31 -04:00 |
|
Wing Lian
|
effbbf6dd1
|
more pruning
|
2023-06-11 10:38:24 -04:00 |
|
Wing Lian
|
c530e4b9c8
|
more config pruning and migrating
|
2023-06-11 10:09:05 -04:00 |
|
Wing Lian
|
77762a5d6b
|
get rid of some configs, formalize pythioa lora config
|
2023-06-11 09:41:41 -04:00 |
|
Wing Lian
|
0c6f928601
|
address PR feedback
|
2023-06-10 14:23:56 -04:00 |
|
Wing Lian
|
1db46a9c72
|
linting fix
|
2023-06-10 14:23:56 -04:00 |
|
Wing Lian
|
39619028a3
|
use pythia-12b, neox-20b is flaky
|
2023-06-10 14:22:30 -04:00 |
|
NanoCode012
|
c8242de725
|
Merge pull request #132 from utensil/falcon-7b-qlora
Axolotl supports falcon + qlora
|
2023-06-09 01:14:03 +09:00 |
|
Utensil
|
79a8f52181
|
Trim trailing whitespace
|
2023-06-08 23:48:57 +08:00 |
|
Utensil
|
a52f4816b0
|
Default wandb_project to empty as suggested
Co-authored-by: NanoCode012 <kevinvong@rocketmail.com>
|
2023-06-08 23:04:19 +08:00 |
|
Utensil
|
c9c050316f
|
Default micro_batch_size to 1 for a safer start
|
2023-06-03 17:26:33 +08:00 |
|
Utensil
|
ca11ae9689
|
Add comments/alternatives for falcon-qlora configs
|
2023-06-03 15:04:02 +08:00 |
|
Utensil
|
fb3d40f197
|
falcon + qlora + xformer mbs 40 gas 2 on A6000
|
2023-06-01 18:29:20 +08:00 |
|
Utensil
|
72bf8aafb6
|
Create config-7b-qlora.yml
|
2023-06-01 00:00:37 +08:00 |
|
Wing Lian
|
c2a0792680
|
swap batch size for gradient accumulation steps to decouple from num gpu
|
2023-05-31 09:38:12 -04:00 |
|
Wing Lian
|
4df9da74e3
|
Merge pull request #105 from viktoriussuwandi/viktoriussuwandi-patch
Viktoriussuwandi patch
|
2023-05-30 15:05:23 -04:00 |
|
Wing Lian
|
2531ea24c1
|
Merge pull request #106 from fearnworks/qlora-openllama-3b-example
Qlora openllama 3b example
|
2023-05-30 15:05:05 -04:00 |
|
NanoCode012
|
392dfd9b07
|
Lint and format
|
2023-05-31 02:53:22 +09:00 |
|
jphillips
|
6cee881d64
|
Update examples/qlora-openllama-3b/README.md
Co-authored-by: Wing Lian <wing.lian@gmail.com>
|
2023-05-30 09:33:33 -05:00 |
|
jphillips
|
ac85c0ed36
|
Add Readme, Clean up comments
|
2023-05-29 14:35:58 -05:00 |
|
jphillips
|
370d057096
|
Add qlora-openllama-3b example
|
2023-05-29 09:07:46 -05:00 |
|
Viktorius Suwandi
|
15e57ba6ee
|
Update wandb_log_model on config.yml
|
2023-05-29 16:33:20 +07:00 |
|
Viktorius Suwandi
|
4eb68ac3f7
|
Update wandb_log_model on config-3b.yml
|
2023-05-29 16:32:49 +07:00 |
|
Viktorius Suwandi
|
fad06befee
|
Update wandb_log_model on config.yml
|
2023-05-29 15:42:38 +07:00 |
|
Wing Lian
|
147241ca66
|
Merge branch 'main' into refactor/rename-4b-to-gptq
|
2023-05-27 09:37:52 -04:00 |
|
Wing Lian
|
157420df13
|
sane starter defaults and add lora
|
2023-05-27 09:33:14 -04:00 |
|
Wing Lian
|
d5f944ce2a
|
add example for falcoln support
|
2023-05-27 09:16:43 -04:00 |
|
Thytu
|
7cf07fc8b3
|
refactor(example): rename 4bit-lora-7b by gptq-lora-7b
Signed-off-by: Thytu <vdmatos@gladia.io>
|
2023-05-27 12:37:53 +00:00 |
|
Thytu
|
dd0065773a
|
refactor(param): rename load_4bit config param by gptq
Signed-off-by: Thytu <vdmatos@gladia.io>
|
2023-05-27 12:36:03 +00:00 |
|