From bc18c1ef6014aedc3f04f26febe3949f9d665cae Mon Sep 17 00:00:00 2001 From: Quarto GHA Workflow Runner Date: Tue, 3 Dec 2024 13:59:16 +0000 Subject: [PATCH] Built site for gh-pages --- .nojekyll | 2 +- docs/dataset-formats/index.html | 10 +++--- index.html | 2 +- search.json | 2 +- sitemap.xml | 56 ++++++++++++++++----------------- 5 files changed, 36 insertions(+), 36 deletions(-) diff --git a/.nojekyll b/.nojekyll index 6c2d63ebf..51fda9628 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -5cbfb703 \ No newline at end of file +64d07371 \ No newline at end of file diff --git a/docs/dataset-formats/index.html b/docs/dataset-formats/index.html index 027b03605..a3ba2dcf1 100644 --- a/docs/dataset-formats/index.html +++ b/docs/dataset-formats/index.html @@ -363,7 +363,7 @@ Description - + Pre-training @@ -371,7 +371,7 @@ Description Data format for a pre-training completion task. - + Instruction Tuning @@ -379,7 +379,7 @@ Description Instruction tuning formats for supervised fine-tuning. - + Conversation @@ -387,7 +387,7 @@ Description Conversation format for supervised fine-tuning. - + Template-Free @@ -395,7 +395,7 @@ Description Construct prompts without a template. - + Custom Pre-Tokenized Dataset diff --git a/index.html b/index.html index 0a187fcaa..0179fefbf 100644 --- a/index.html +++ b/index.html @@ -568,7 +568,7 @@ pre > code.sourceCode > span > a:first-child::before { text-decoration: underlin

Usage

# preprocess datasets - optional but recommended
-CUDA_VISIBLE_DEVICES="" python -m axolotl.cli.preprocess examples/openllama-3b/lora.yml
+CUDA_VISIBLE_DEVICES="0" python -m axolotl.cli.preprocess examples/openllama-3b/lora.yml
 
 # finetune lora
 accelerate launch -m axolotl.cli.train examples/openllama-3b/lora.yml
diff --git a/search.json b/search.json
index 67f5332d1..2d6e7b9c0 100644
--- a/search.json
+++ b/search.json
@@ -24,7 +24,7 @@
     "href": "index.html#quickstart",
     "title": "Axolotl",
     "section": "Quickstart ⚡",
-    "text": "Quickstart ⚡\nGet started with Axolotl in just a few steps! This quickstart guide will walk you through setting up and running a basic fine-tuning task.\nRequirements: Nvidia GPU (Ampere architecture or newer for bf16 and Flash Attention), Python >=3.10 and PyTorch >=2.3.1.\ngit clone https://github.com/axolotl-ai-cloud/axolotl\ncd axolotl\n\npip3 install packaging ninja\npip3 install -e '.[flash-attn,deepspeed]'\n\nUsage\n# preprocess datasets - optional but recommended\nCUDA_VISIBLE_DEVICES=\"\" python -m axolotl.cli.preprocess examples/openllama-3b/lora.yml\n\n# finetune lora\naccelerate launch -m axolotl.cli.train examples/openllama-3b/lora.yml\n\n# inference\naccelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \\\n    --lora_model_dir=\"./outputs/lora-out\"\n\n# gradio\naccelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \\\n    --lora_model_dir=\"./outputs/lora-out\" --gradio\n\n# remote yaml files - the yaml config can be hosted on a public URL\n# Note: the yaml config must directly link to the **raw** yaml\naccelerate launch -m axolotl.cli.train https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/examples/openllama-3b/lora.yml",
+    "text": "Quickstart ⚡\nGet started with Axolotl in just a few steps! This quickstart guide will walk you through setting up and running a basic fine-tuning task.\nRequirements: Nvidia GPU (Ampere architecture or newer for bf16 and Flash Attention), Python >=3.10 and PyTorch >=2.3.1.\ngit clone https://github.com/axolotl-ai-cloud/axolotl\ncd axolotl\n\npip3 install packaging ninja\npip3 install -e '.[flash-attn,deepspeed]'\n\nUsage\n# preprocess datasets - optional but recommended\nCUDA_VISIBLE_DEVICES=\"0\" python -m axolotl.cli.preprocess examples/openllama-3b/lora.yml\n\n# finetune lora\naccelerate launch -m axolotl.cli.train examples/openllama-3b/lora.yml\n\n# inference\naccelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \\\n    --lora_model_dir=\"./outputs/lora-out\"\n\n# gradio\naccelerate launch -m axolotl.cli.inference examples/openllama-3b/lora.yml \\\n    --lora_model_dir=\"./outputs/lora-out\" --gradio\n\n# remote yaml files - the yaml config can be hosted on a public URL\n# Note: the yaml config must directly link to the **raw** yaml\naccelerate launch -m axolotl.cli.train https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/examples/openllama-3b/lora.yml",
     "crumbs": [
       "Home"
     ]
diff --git a/sitemap.xml b/sitemap.xml
index 0539a9984..0dadb2448 100644
--- a/sitemap.xml
+++ b/sitemap.xml
@@ -2,114 +2,114 @@
 
   
     https://axolotl-ai-cloud.github.io/axolotl/index.html
-    2024-12-03T13:22:35.000Z
+    2024-12-03T13:58:34.234Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/LICENSE.html
-    2024-12-03T13:22:35.004Z
+    2024-12-03T13:58:34.238Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/nccl.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/input_output.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset_preprocessing.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/torchao.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/rlhf.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/config.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/template_free.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/conversation.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/tokenized.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/mac.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multi-node.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/FAQS.html
-    2024-12-03T13:22:34.984Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/TODO.html
-    2024-12-03T13:22:34.984Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/faq.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/debugging.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/inst_tune.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/pretraining.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/dataset-formats/index.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/unsloth.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multimodal.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/batch_vs_grad.html
-    2024-12-03T13:22:34.984Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/fsdp_qlora.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/multipack.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/docs/amd_hpc.html
-    2024-12-03T13:22:34.984Z
+    2024-12-03T13:58:34.222Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/examples/colab-notebooks/colab-axolotl-example.html
-    2024-12-03T13:22:34.988Z
+    2024-12-03T13:58:34.226Z
   
   
     https://axolotl-ai-cloud.github.io/axolotl/src/axolotl/integrations/cut_cross_entropy/ACKNOWLEDGEMENTS.html
-    2024-12-03T13:22:35.004Z
+    2024-12-03T13:58:34.238Z