From b502392e828d25069067e041dfe08030811ac436 Mon Sep 17 00:00:00 2001 From: Hamel Husain Date: Thu, 11 Jan 2024 16:41:58 -0800 Subject: [PATCH] Update README.md (#1103) * Update README.md * Update README.md --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 7b303fa78..c9731ad70 100644 --- a/README.md +++ b/README.md @@ -139,7 +139,7 @@ accelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \ A more powerful Docker command to run would be this: ```bash - docker run --privileged --gpus '"all"' --shm-size 10g --rm -it --name axolotl --ipc=host --ulimit memlock=-1 --ulimit stack=67108864 --mount type=volume,src=axolotl,target=/workspace/axolotl -v ${HOME}/.cache/huggingface:/root/.cache/huggingface winglian/axolotl:main-py3.10-cu118-2.0.1 +docker run --privileged --gpus '"all"' --shm-size 10g --rm -it --name axolotl --ipc=host --ulimit memlock=-1 --ulimit stack=67108864 --mount type=bind,src="${PWD}",target=/workspace/axolotl -v ${HOME}/.cache/huggingface:/root/.cache/huggingface winglian/axolotl:main-py3.10-cu118-2.0.1 ``` It additionally: