chore(doc): add explanation on fsdp_transformer_layer_cls_to_wrap (#2429) [skip ci]

This commit is contained in:
NanoCode012
2025-03-21 22:59:22 +07:00
committed by GitHub
parent c907ac173e
commit 61825a464a

View File

@@ -37,6 +37,10 @@ description: Frequently asked questions
> A: Yes, since Axolotl is just Python, please see `src/axolotl/cli/main.py` on how each command is called.
**Q: How to know the value to use for `fsdp_transformer_layer_cls_to_wrap`?**
> A: This is the class name of the transformer layer to wrap with FSDP. For example, for `LlamaForCausalLM`, the value is `LlamaDecoderLayer`. To find this for a specific model, check the model's `PreTrainedModel` definition and look for `_no_split_modules` variable in the `modeling_<model_name>.py` file within `transformers` library.
### Chat templates
**Q: `jinja2.exceptions.UndefinedError: 'dict object' has no attribute 'content' / 'role' / ____`**