* various batch of fixes * more tweaks * fix autoawq requirement for torch flexibility * simplify conditionals * multi-node fixes wip * bump transformers and include 405b qlora+fsdp yaml
47 lines
714 B
Plaintext
47 lines
714 B
Plaintext
--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
|
|
packaging==23.2
|
|
peft==0.11.1
|
|
transformers==4.43.3
|
|
tokenizers==0.19.1
|
|
bitsandbytes==0.43.1
|
|
accelerate==0.32.0
|
|
deepspeed==0.14.4
|
|
pydantic==2.6.3
|
|
addict
|
|
fire
|
|
PyYAML>=6.0
|
|
requests
|
|
datasets==2.19.1
|
|
flash-attn==2.6.2
|
|
sentencepiece
|
|
wandb
|
|
einops
|
|
xformers==0.0.27
|
|
optimum==1.16.2
|
|
hf_transfer
|
|
colorama
|
|
numba
|
|
numpy>=1.24.4
|
|
# qlora things
|
|
evaluate==0.4.1
|
|
scipy
|
|
scikit-learn==1.2.2
|
|
pynvml
|
|
art
|
|
fschat @ git+https://github.com/lm-sys/FastChat.git@27a05b04a35510afb1d767ae7e5990cbd278f8fe
|
|
gradio==3.50.2
|
|
tensorboard
|
|
python-dotenv==1.0.1
|
|
autoawq>=0.2.5
|
|
|
|
mamba-ssm==1.2.0.post1
|
|
|
|
# remote filesystems
|
|
s3fs
|
|
gcsfs
|
|
# adlfs
|
|
|
|
trl==0.9.6
|
|
zstandard==0.22.0
|
|
fastcore
|