* bump hf deps * upgrade liger-kernel too * install cce from fork for transformers fix * fix reference to vocab size in gemma3 patch * use padding_idx instead of pad_token_id * remove fixed gemma3 patch * use updated cce fork * fix local mllama cce patches w docstring * add test for multipack with trainer setup and fix trainer for trainer refactor upstream * bump modal version * guard for iterable datasetS * mllama model arch layout changed in latest transformers * fix batch sampler with drop_last * fix: address upstream vlm changes for lora * fix: update references to old lora target path * fix: remove mllama fa2 patch due to upstream fix * fix: lora kernel patch path for multimodal models * fix: removed mllama from quarto * run test for came optim on 2.6.0+ * fix fsdp2 patch and remove deprecated patch * make sure to set sequence_parallel_degree for grpo * Add SP test for GRPO * add sp to grpo config for trainer * use reward_funcs as kwarg to grpo trainer * fix the comprehension for reward funcs * reward funcs already passed in as args * init sp_group right before training * fix check for adding models to SP context * make sure to pass args to super * upgrade deepspeed * use updated trl and add reasoning flags for vllm * patch the worker --------- Co-authored-by: NanoCode012 <nano@axolotl.ai>
71 lines
1.0 KiB
Plaintext
71 lines
1.0 KiB
Plaintext
--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
|
|
|
|
# START section of dependencies that don't install on Darwin/MacOS
|
|
bitsandbytes==0.45.4
|
|
triton>=3.0.0
|
|
mamba-ssm==1.2.0.post1
|
|
xformers>=0.0.23.post1
|
|
autoawq==0.2.7.post3
|
|
liger-kernel==0.5.10
|
|
# END section
|
|
|
|
packaging==23.2
|
|
|
|
huggingface_hub==0.32.2
|
|
peft==0.15.2
|
|
transformers==4.52.3
|
|
tokenizers>=0.21.1
|
|
accelerate==1.7.0
|
|
datasets==3.6.0
|
|
deepspeed>=0.17.0
|
|
trl==0.18.1
|
|
hf_xet==1.1.2
|
|
hqq==0.2.5
|
|
|
|
optimum==1.16.2
|
|
hf_transfer
|
|
sentencepiece
|
|
gradio==5.23.3
|
|
|
|
modal==0.70.5
|
|
pydantic==2.10.6
|
|
addict
|
|
fire
|
|
PyYAML>=6.0
|
|
requests
|
|
wandb
|
|
einops
|
|
colorama
|
|
numba
|
|
numpy>=1.24.4,<=2.0.1
|
|
|
|
# qlora things
|
|
evaluate==0.4.1
|
|
scipy
|
|
scikit-learn==1.4.2
|
|
nvidia-ml-py==12.560.30
|
|
art
|
|
tensorboard
|
|
python-dotenv==1.0.1
|
|
|
|
# remote filesystems
|
|
s3fs>=2024.5.0
|
|
gcsfs>=2024.5.0
|
|
adlfs>=2024.5.0
|
|
ocifs==1.3.2
|
|
|
|
zstandard==0.22.0
|
|
fastcore
|
|
|
|
# lm eval harness
|
|
lm_eval==0.4.7
|
|
langdetect==1.0.9
|
|
immutabledict==4.2.0
|
|
antlr4-python3-runtime==4.13.2
|
|
|
|
torchao==0.10.0
|
|
schedulefree==1.4.1
|
|
|
|
axolotl-contribs-lgpl==0.0.6
|
|
axolotl-contribs-mit==0.0.3
|