From b06d3e364554f84be3435b8f41084ca88303a30a Mon Sep 17 00:00:00 2001 From: Wing Lian Date: Thu, 20 Jul 2023 01:02:08 -0400 Subject: [PATCH] explicitly pin flash attention 1 to v1.0.9 --- docker/Dockerfile-base | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/docker/Dockerfile-base b/docker/Dockerfile-base index adf7996ee..ff60ca504 100644 --- a/docker/Dockerfile-base +++ b/docker/Dockerfile-base @@ -38,8 +38,9 @@ WORKDIR /workspace ARG TORCH_CUDA_ARCH_LIST="7.0 7.5 8.0 8.6+PTX" -RUN git clone https://github.com/HazyResearch/flash-attention.git && \ +RUN git clone https://github.com/Dao-AILab/flash-attention.git && \ cd flash-attention && \ + git checkout v1.0.9 && \ python3 setup.py bdist_wheel && \ cd csrc/fused_dense_lib && \ python3 setup.py bdist_wheel && \