Logo
Explore Help
Sign In
tocmo0nlord/axolotl
1
0
Fork 0
You've already forked axolotl
Code Issues Pull Requests Actions 3 Packages Projects Releases Wiki Activity
Files
1f09f48d8fa5f244aef05461b0b2d557860875da
axolotl/examples/llama-3
History
Aaditya Ura (looking for PhD Fall’24) cb023c70db Update instruct-lora-8b.yml (#1789) [skip ci]
Config is giving an error if not using the end of the token as the `pad_to_sequence_len` is true.
2024-08-05 12:43:20 -04:00
..
fft-8b.yaml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
instruct-dpo-lora-8b.yml
Add a chat_template prompt strategy for DPO (#1725)
2024-07-21 09:10:42 -04:00
instruct-lora-8b.yml
Update instruct-lora-8b.yml (#1789) [skip ci]
2024-08-05 12:43:20 -04:00
lora-8b.yml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
qlora-fsdp-70b.yaml
update outputs path so that we can mount workspace to /workspace/data (#1623)
2024-05-15 12:44:13 -04:00
qlora-fsdp-405b.yaml
qlora-fsdp ram efficient loading with hf trainer (#1791)
2024-07-30 19:21:38 -04:00
qlora.yml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
README.md
llama-3 examples (#1537)
2024-04-18 14:28:03 -04:00

README.md

Llama-3

https://llama.meta.com/llama3/

8B Base Model

  • Full Fine Tune
    • Single GPU @ 48GB VRAM
  • LoRA
    • Single GPU @ 11GB VRAM

70B Base Model

  • QLORA+FSDP
    • Dual GPU @ 21GB VRAM
Reference in New Issue View Git Blame Copy Permalink
Powered by Gitea Version: 1.25.4 Page: 18ms Template: 2ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API