fix: checkpoint saving with deepspeed (#1321)

This commit is contained in:
NanoCode012
2024-02-27 15:46:44 +09:00
committed by GitHub
parent 0f6af36d50
commit 5be8b555a0
2 changed files with 2 additions and 2 deletions

View File

@@ -5,7 +5,7 @@ transformers @ git+https://github.com/huggingface/transformers.git@ae49b218c3d71
tokenizers==0.15.0
bitsandbytes>=0.41.1
accelerate==0.26.1
deepspeed>=0.13.1
deepspeed==0.13.1
pydantic>=2.5.3
addict
fire

View File

@@ -74,7 +74,7 @@ setup(
"fused-dense-lib @ git+https://github.com/Dao-AILab/flash-attention@v2.3.3#subdirectory=csrc/fused_dense_lib",
],
"deepspeed": [
"deepspeed>=0.13.1",
"deepspeed==0.13.1",
"deepspeed-kernels",
],
"mamba-ssm": [