diff --git a/.nojekyll b/.nojekyll index 09ce6e542..73268f02f 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -756ab801 \ No newline at end of file +3ace7c03 \ No newline at end of file diff --git a/docs/custom_integrations.html b/docs/custom_integrations.html index 45fb54ec6..e8c9dac3e 100644 --- a/docs/custom_integrations.html +++ b/docs/custom_integrations.html @@ -963,7 +963,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true}); -
pip3 uninstall -y cut-cross-entropy && pip3 install "cut-cross-entropy[transformers] @ git+https://github.com/axolotl-ai-cloud/ml-cross-entropy.git@e8ad129"
+
pip3 uninstall -y cut-cross-entropy && pip3 install "cut-cross-entropy[transformers] @ git+https://github.com/axolotl-ai-cloud/ml-cross-entropy.git@fa9a7fe"

Usage

@@ -1015,8 +1015,10 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
  • ministral3
  • mistral
  • mistral3
  • +
  • mistral4
  • mixtral
  • mllama
  • +
  • nemotron_h
  • olmo
  • olmo2
  • olmo3
  • diff --git a/docs/multimodal.html b/docs/multimodal.html index 21a87e37a..0e8b6cadd 100644 --- a/docs/multimodal.html +++ b/docs/multimodal.html @@ -762,6 +762,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
  • Pixtral
  • Llava-1.5
  • Mistral-Small-3.1
  • +
  • Mistral-Small-4
  • Magistral-Small-2509
  • Voxtral
  • Gemma-3
  • @@ -815,6 +816,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
  • Pixtral
  • Llava-1.5
  • Mistral-Small-3.1
  • +
  • Mistral-Small-4
  • Magistral-Small-2509
  • Voxtral
  • Gemma-3
  • @@ -922,6 +924,10 @@ Tip
    base_model: mistralai/Mistral-Small-3.1-24B-Instruct-2503
    +
    +

    Mistral-Small-4

    +
    base_model: mistralai/Mistral-Small-4-119B-2603
    +

    Magistral-Small-2509

    @@ -937,7 +943,7 @@ Tip

    Please make sure to install vision lib via pip install 'mistral-common[opencv]==1.8.5'

    -
    base_model: mistralai/Magistral-Small-2509
    +
    base_model: mistralai/Magistral-Small-2509

    Voxtral

    @@ -954,9 +960,9 @@ Tip

    Please make sure to install audio lib via pip3 install librosa==0.11.0 'mistral_common[audio]==1.8.3'

    -
    base_model: mistralai/Voxtral-Mini-3B-2507
    -
    -processor_type: VoxtralProcessor
    +
    base_model: mistralai/Voxtral-Mini-3B-2507
    +
    +processor_type: VoxtralProcessor

    Gemma-3

    @@ -974,9 +980,9 @@ Tip

    For multi-modal 4B/12B/27B models, use the following config:

    -
    base_model: google/gemma-3-4b-it
    -
    -chat_template: gemma3
    +
    base_model: google/gemma-3-4b-it
    +
    +chat_template: gemma3

    Gemma-3n

    @@ -1006,36 +1012,36 @@ Tip

    Please make sure to install timm via pip3 install timm==1.0.17

    -
    base_model: google/gemma-3n-E2B-it
    -
    -chat_template: gemma3n
    +
    base_model: google/gemma-3n-E2B-it
    +
    +chat_template: gemma3n

    Qwen2-VL

    -
    base_model: Qwen/Qwen2-VL-7B-Instruct
    -
    -chat_template: qwen2_vl
    +
    base_model: Qwen/Qwen2-VL-7B-Instruct
    +
    +chat_template: qwen2_vl

    Qwen2.5-VL

    -
    base_model: Qwen/Qwen2.5-VL-7B-Instruct
    -
    -chat_template: qwen2_vl  # same as qwen2-vl
    +
    base_model: Qwen/Qwen2.5-VL-7B-Instruct
    +
    +chat_template: qwen2_vl  # same as qwen2-vl

    Qwen3-VL

    -
    base_model: Qwen/Qwen3-VL-4B-Instruct
    -
    -chat_template: qwen2_vl  # same as qwen2-vl
    +
    base_model: Qwen/Qwen3-VL-4B-Instruct
    +
    +chat_template: qwen2_vl  # same as qwen2-vl

    GLM-4.6V

    Both GLM-4.6V (106B MoE) and GLM-4.6V-Flash (9B) are supported.

    -
    # GLM-4.6V (106B MoE version)
    -base_model: zai-org/GLM-4.6V
    -
    -# OR GLM-4.6V-Flash (9B version)
    -base_model: zai-org/GLM-4.6V-Flash
    +
    # GLM-4.6V (106B MoE version)
    +base_model: zai-org/GLM-4.6V
    +
    +# OR GLM-4.6V-Flash (9B version)
    +base_model: zai-org/GLM-4.6V-Flash

    SmolVLM2

    @@ -1052,7 +1058,7 @@ Tip

    Please make sure to install num2words via pip3 install num2words==0.5.14

    -
    base_model: HuggingFaceTB/SmolVLM2-500M-Video-Instruct
    +
    base_model: HuggingFaceTB/SmolVLM2-500M-Video-Instruct

    LFM2-VL

    @@ -1069,7 +1075,7 @@ Warning

    Please uninstall causal-conv1d via pip3 uninstall -y causal-conv1d

    -
    base_model: LiquidAI/LFM2-VL-450M
    +
    base_model: LiquidAI/LFM2-VL-450M

    Intern-VL

    @@ -1086,7 +1092,7 @@ Tip

    Please make sure to install timm via pip3 install timm==1.0.19

    -
    base_model: OpenGVLab/InternVL3_5-8B
    +
    base_model: OpenGVLab/InternVL3_5-8B
    @@ -1171,31 +1177,31 @@ Warning

    Example

    Here is an example of a multi-modal dataset:

    -
    [
    -  {
    -    "messages": [
    -        {
    -            "role": "system",
    -            "content": [
    -              {"type": "text", "text": "You are a helpful assistant."}
    -              ]
    -        },
    -        {
    -            "role": "user",
    -            "content": [
    -                {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg"},
    -                {"type": "text", "text": "Describe this image in detail."}
    -            ]
    -        },
    -        {
    -            "role": "assistant",
    -            "content": [
    -              {"type": "text", "text": "The image is a bee."}
    -            ]
    -        }
    -    ]
    -  }
    -]
    +
    [
    +  {
    +    "messages": [
    +        {
    +            "role": "system",
    +            "content": [
    +              {"type": "text", "text": "You are a helpful assistant."}
    +              ]
    +        },
    +        {
    +            "role": "user",
    +            "content": [
    +                {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg"},
    +                {"type": "text", "text": "Describe this image in detail."}
    +            ]
    +        },
    +        {
    +            "role": "assistant",
    +            "content": [
    +              {"type": "text", "text": "The image is a bee."}
    +            ]
    +        }
    +    ]
    +  }
    +]
    diff --git a/examples/colab-notebooks/colab-axolotl-example.html b/examples/colab-notebooks/colab-axolotl-example.html index 92cfa3323..1eacd9278 100644 --- a/examples/colab-notebooks/colab-axolotl-example.html +++ b/examples/colab-notebooks/colab-axolotl-example.html @@ -810,7 +810,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
    %%capture
     # This step can take ~5-10 minutes to install dependencies
     !pip install --no-build-isolation axolotl[flash-attn]>=0.9.1
    -!pip install "cut-cross-entropy[transformers] @ git+https://github.com/axolotl-ai-cloud/ml-cross-entropy.git@e8ad129"
    +!pip install "cut-cross-entropy[transformers] @ git+https://github.com/axolotl-ai-cloud/ml-cross-entropy.git@fa9a7fe"

    Demo: Talk Like a Pirate

    diff --git a/index.html b/index.html index 52b106a0a..2090ffb38 100644 --- a/index.html +++ b/index.html @@ -809,7 +809,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});