* [llama4] fix the mm yaml, add scout single gpu yaml * add README for llama4 * rename to specify fsdp
Llama 4 by Meta AI
Available Examples
Llama 4 Scout 17Bx16Experts (109B)
Our Single GPU implementation for Llama 4 Scout uses only 68.5GB VRAM for post-training with 4k context length @ 546 tokens/second.