Logo
Explore Help
Sign In
tocmo0nlord/axolotl
1
0
Fork 0
You've already forked axolotl
Code Issues Pull Requests Actions 3 Packages Projects Releases Wiki Activity
Files
2b890ead05dcab97f2d0f15f6f9891ba93b8a577
axolotl/examples/llama-3
History
Aaditya Ura (looking for PhD Fall’24) cb023c70db Update instruct-lora-8b.yml (#1789) [skip ci]
Config is giving an error if not using the end of the token as the `pad_to_sequence_len` is true.
2024-08-05 12:43:20 -04:00
..
fft-8b.yaml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
instruct-dpo-lora-8b.yml
Add a chat_template prompt strategy for DPO (#1725)
2024-07-21 09:10:42 -04:00
instruct-lora-8b.yml
Update instruct-lora-8b.yml (#1789) [skip ci]
2024-08-05 12:43:20 -04:00
lora-8b.yml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
qlora-fsdp-70b.yaml
update outputs path so that we can mount workspace to /workspace/data (#1623)
2024-05-15 12:44:13 -04:00
qlora-fsdp-405b.yaml
qlora-fsdp ram efficient loading with hf trainer (#1791)
2024-07-30 19:21:38 -04:00
qlora.yml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
README.md
llama-3 examples (#1537)
2024-04-18 14:28:03 -04:00

README.md

Llama-3

https://llama.meta.com/llama3/

8B Base Model

  • Full Fine Tune
    • Single GPU @ 48GB VRAM
  • LoRA
    • Single GPU @ 11GB VRAM

70B Base Model

  • QLORA+FSDP
    • Dual GPU @ 21GB VRAM
Reference in New Issue View Git Blame Copy Permalink
Powered by Gitea Version: 1.25.4 Page: 227ms Template: 1ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API