Files
axolotl/examples/llama-4
Wing Lian bf9efe2a09 [llama4] fix the mm yaml, add scout single gpu yaml (#2510)
* [llama4] fix the mm yaml, add scout single gpu yaml

* add README for llama4

* rename to specify fsdp
2025-04-09 02:52:45 -04:00
..

Llama 4 by Meta AI

Available Examples

Llama 4 Scout 17Bx16Experts (109B)

Our Single GPU implementation for Llama 4 Scout uses only 68.5GB VRAM for post-training with 4k context length @ 546 tokens/second.