Logo
Explore Help
Sign In
tocmo0nlord/axolotl
1
0
Fork 0
You've already forked axolotl
Code Issues Pull Requests Actions 3 Packages Projects Releases Wiki Activity
Files
cb023c70dbbbe611cba746375b4cb72b8a598b3b
axolotl/examples/llama-3
History
Aaditya Ura (looking for PhD Fall’24) cb023c70db Update instruct-lora-8b.yml (#1789) [skip ci]
Config is giving an error if not using the end of the token as the `pad_to_sequence_len` is true.
2024-08-05 12:43:20 -04:00
..
fft-8b.yaml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
instruct-dpo-lora-8b.yml
Add a chat_template prompt strategy for DPO (#1725)
2024-07-21 09:10:42 -04:00
instruct-lora-8b.yml
Update instruct-lora-8b.yml (#1789) [skip ci]
2024-08-05 12:43:20 -04:00
lora-8b.yml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
qlora-fsdp-70b.yaml
update outputs path so that we can mount workspace to /workspace/data (#1623)
2024-05-15 12:44:13 -04:00
qlora-fsdp-405b.yaml
qlora-fsdp ram efficient loading with hf trainer (#1791)
2024-07-30 19:21:38 -04:00
qlora.yml
bump transformers and set roundup_power2_divisions for more VRAM improvements, low bit ao optimizers (#1769)
2024-07-19 00:47:07 -04:00
README.md
llama-3 examples (#1537)
2024-04-18 14:28:03 -04:00

README.md

Llama-3

https://llama.meta.com/llama3/

8B Base Model

  • Full Fine Tune
    • Single GPU @ 48GB VRAM
  • LoRA
    • Single GPU @ 11GB VRAM

70B Base Model

  • QLORA+FSDP
    • Dual GPU @ 21GB VRAM
Reference in New Issue View Git Blame Copy Permalink
Powered by Gitea Version: 1.25.4 Page: 155ms Template: 1ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API