* auto gptq support * more tweaks and add yml * remove old gptq docker * don't need explicit peft install for tests * fix setup.py to use extra index url install torch for tests fix cuda version for autogptq index set torch in requirements so that it installs properly move gptq install around to work with github cicd * gptq doesn't play well with sample packing * address pr feedback * remove torch install for now * set quantization_config from model config * Fix the implementation for getting quant config from model config
30 lines
839 B
Docker
30 lines
839 B
Docker
ARG BASE_TAG=main-base
|
|
FROM winglian/axolotl-base:$BASE_TAG
|
|
|
|
ARG TORCH_CUDA_ARCH_LIST="7.0 7.5 8.0 8.6+PTX"
|
|
ARG AXOLOTL_EXTRAS=""
|
|
ARG CUDA="118"
|
|
ENV BNB_CUDA_VERSION=$CUDA
|
|
|
|
RUN apt-get update && \
|
|
apt-get install -y vim curl
|
|
|
|
WORKDIR /workspace
|
|
|
|
RUN git clone --depth=1 https://github.com/OpenAccess-AI-Collective/axolotl.git
|
|
# If AXOLOTL_EXTRAS is set, append it in brackets
|
|
RUN cd axolotl && \
|
|
if [ "$AXOLOTL_EXTRAS" != "" ] ; then \
|
|
pip install -e .[flash-attn,gptq,$AXOLOTL_EXTRAS]; \
|
|
else \
|
|
pip install -e .[flash-attn,gptq]; \
|
|
fi
|
|
|
|
# fix so that git fetch/pull from remote works
|
|
RUN cd axolotl && \
|
|
git config remote.origin.fetch "+refs/heads/*:refs/remotes/origin/*" && \
|
|
git config --get remote.origin.fetch
|
|
|
|
# helper for huggingface-login cli
|
|
RUN git config --global credential.helper store
|