* import deepspeed integration * monkeypatch peft adapater with deepspeed for resume from checkpoint * fix patch * fix patches attempt 2 * make sure to set lora_model_dir * skip pylint for deepspeed.utils * pick up upstream fix in transformers * remove monkeypatch for deepspeed/peft fix * no need to set the lora_model_dir on resume * unset load_in_*bit when using quant config * guard before del * better handling of load_in* kwargs
42 lines
674 B
Plaintext
42 lines
674 B
Plaintext
--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
|
|
packaging==23.2
|
|
peft @ git+https://github.com/huggingface/peft.git
|
|
transformers @ git+https://github.com/huggingface/transformers.git@bebeeee01275c32fccec3fa36d8b148d3813a7dc
|
|
tokenizers==0.15.0
|
|
bitsandbytes>=0.41.1
|
|
accelerate==0.26.1
|
|
deepspeed>=0.13.1
|
|
addict
|
|
fire
|
|
PyYAML>=6.0
|
|
datasets>=2.15.0
|
|
flash-attn==2.3.3
|
|
sentencepiece
|
|
wandb
|
|
einops
|
|
xformers==0.0.22
|
|
optimum==1.16.2
|
|
hf_transfer
|
|
colorama
|
|
numba
|
|
numpy>=1.24.4
|
|
mlflow
|
|
# qlora things
|
|
evaluate==0.4.0
|
|
scipy
|
|
scikit-learn==1.2.2
|
|
pynvml
|
|
art
|
|
fschat==0.2.34
|
|
gradio==3.50.2
|
|
tensorboard
|
|
|
|
mamba-ssm==1.1.1
|
|
|
|
# remote filesystems
|
|
s3fs
|
|
gcsfs
|
|
# adlfs
|
|
|
|
trl>=0.7.9
|