* restore to current phi modeling code from phi-2 * enable gradient checkpointing * don't cast everything to float32 all the time * gradient checkpointing for phi2 ParallelBlock module too * fix enabling flash attn for phi2 * add comment about import * fix phi2 example * fix model type check for tokenizer * revert float32 -> bf16 casting changes * support fused dense flash attn * fix the repo for flash-attn * add package name for subdir pkg * fix the data collator when not using sample packing * install packaging for pytests in ci * also fix setup to not install flash attn fused dense subdir if not extras * split out the fused-dense-lib in extra requires * don't train w group_by_length for phi * update integration test to use phi2 * set max steps and save steps for phi e2e tests * try to workaround ssave issue in ci * skip phi2 e2e test for now
43 lines
803 B
Plaintext
43 lines
803 B
Plaintext
--extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
|
|
auto-gptq==0.5.1
|
|
packaging
|
|
peft==0.6.0
|
|
transformers @ git+https://github.com/huggingface/transformers.git@3cefac1d974db5e2825a0cb2b842883a628be7a0
|
|
tokenizers==0.15.0
|
|
bitsandbytes>=0.41.1
|
|
accelerate==0.24.1
|
|
deepspeed
|
|
addict
|
|
fire
|
|
PyYAML>=6.0
|
|
datasets>=2.15.0
|
|
flash-attn==2.3.3
|
|
fused-dense-lib @ git+https://github.com/Dao-AILab/flash-attention@v2.3.3#subdirectory=csrc/fused_dense_lib
|
|
sentencepiece
|
|
wandb
|
|
einops
|
|
xformers==0.0.22
|
|
optimum==1.13.2
|
|
hf_transfer
|
|
colorama
|
|
numba
|
|
numpy>=1.24.4
|
|
# qlora things
|
|
bert-score==0.3.13
|
|
evaluate==0.4.0
|
|
rouge-score==0.1.2
|
|
scipy
|
|
scikit-learn==1.2.2
|
|
pynvml
|
|
art
|
|
fschat==0.2.34
|
|
gradio==3.50.2
|
|
tensorboard
|
|
|
|
# remote filesystems
|
|
s3fs
|
|
gcsfs
|
|
# adlfs
|
|
|
|
trl @ git+https://github.com/huggingface/trl.git@main
|