Files
axolotl/examples/llama-4

Llama 4 by Meta AI

Available Examples

Llama 4 Scout 17Bx16Experts (109B)

Our Single GPU implementation for Llama 4 Scout uses only 68.5GB VRAM for post-training with 4k context length @ 546 tokens/second.