Logo
Explore Help
Sign In
tocmo0nlord/axolotl
1
0
Fork 0
You've already forked axolotl
Code Issues Pull Requests Actions 3 Packages Projects Releases Wiki Activity
2,080 Commits 319 Branches 32 Tags
attention_enum
Commit Graph

161 Commits

Author SHA1 Message Date
Wing Lian
2df63ef815 refactor trainer setup to account for deepspeed integration 2023-04-15 12:16:42 -04:00
Wing Lian
b164725417 improve prepared dataset loading, fix inference 2023-04-15 12:14:52 -04:00
Wing Lian
937f44f021 helpful info output 2023-04-15 00:03:43 -04:00
Wing Lian
902dd0ab47 fix issue with completed model being empty
see https://github.com/huggingface/peft/issues/286#issuecomment-1501617281
2023-04-14 23:57:55 -04:00
Wing Lian
80b2ed29d8 various bugfixes 2023-04-14 21:37:07 -04:00
Wing Lian
45f77dd51e bettter handling of llama model import 2023-04-14 19:30:41 -04:00
Wing Lian
949a27be21 more fixes and prep for llama training 2023-04-14 18:30:09 -04:00
Wing Lian
f2a2029d0d config chooser, update readme instructions, device config, llama flash attention, debug out the labels, fix config key checks, other bugfixes 2023-04-14 12:18:56 -04:00
Wing Lian
a6028d302e black formatting 2023-04-14 07:25:52 -04:00
Wing Lian
8d959a7e26 make it work with pythia in the cloud 2023-04-14 07:24:55 -04:00
Wing Lian
ce24f5e246 WIP for axolotl trainer 2023-04-14 00:20:05 -04:00
First Previous 1 2 3 4 Next Last
Powered by Gitea Version: 1.25.4 Page: 106ms Template: 2ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API