* mxfp4 axo * import lint * test for qat mxfp4 * config for mxfp4 * add qat: * pass base config * MXFakeQuantizeConfig * lint * tune config so it fits in 32GB VRAM --------- Co-authored-by: Wing Lian <wing@axolotl.ai>
* torch_dtype -> dtype * torch_dtype -> dtype
axolotl quantize
* black, isort, flake8 -> ruff * remove unused * add back needed import * fix
QAT and quantization w/torchao