diff --git a/.nojekyll b/.nojekyll index 576a9293d..62ce666e2 100644 --- a/.nojekyll +++ b/.nojekyll @@ -1 +1 @@ -c00becba \ No newline at end of file +0a88f6d8 \ No newline at end of file diff --git a/docs/multimodal.html b/docs/multimodal.html index 0e8b6cadd..c43616d6b 100644 --- a/docs/multimodal.html +++ b/docs/multimodal.html @@ -770,6 +770,7 @@ gtag('config', 'G-9KYCVJBNMQ', { 'anonymize_ip': true});
base_model: Qwen/Qwen3.5-9B
+
+chat_template: qwen3_5Both GLM-4.6V (106B MoE) and GLM-4.6V-Flash (9B) are supported.
-# GLM-4.6V (106B MoE version)
-base_model: zai-org/GLM-4.6V
-
-# OR GLM-4.6V-Flash (9B version)
-base_model: zai-org/GLM-4.6V-Flash# GLM-4.6V (106B MoE version)
+base_model: zai-org/GLM-4.6V
+
+# OR GLM-4.6V-Flash (9B version)
+base_model: zai-org/GLM-4.6V-FlashPlease make sure to install num2words via pip3 install num2words==0.5.14
base_model: HuggingFaceTB/SmolVLM2-500M-Video-Instructbase_model: HuggingFaceTB/SmolVLM2-500M-Video-InstructPlease uninstall causal-conv1d via pip3 uninstall -y causal-conv1d
base_model: LiquidAI/LFM2-VL-450Mbase_model: LiquidAI/LFM2-VL-450MPlease make sure to install timm via pip3 install timm==1.0.19
base_model: OpenGVLab/InternVL3_5-8Bbase_model: OpenGVLab/InternVL3_5-8BHere is an example of a multi-modal dataset:
-[
- {
- "messages": [
- {
- "role": "system",
- "content": [
- {"type": "text", "text": "You are a helpful assistant."}
- ]
- },
- {
- "role": "user",
- "content": [
- {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg"},
- {"type": "text", "text": "Describe this image in detail."}
- ]
- },
- {
- "role": "assistant",
- "content": [
- {"type": "text", "text": "The image is a bee."}
- ]
- }
- ]
- }
-][
+ {
+ "messages": [
+ {
+ "role": "system",
+ "content": [
+ {"type": "text", "text": "You are a helpful assistant."}
+ ]
+ },
+ {
+ "role": "user",
+ "content": [
+ {"type": "image", "url": "https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg"},
+ {"type": "text", "text": "Describe this image in detail."}
+ ]
+ },
+ {
+ "role": "assistant",
+ "content": [
+ {"type": "text", "text": "The image is a bee."}
+ ]
+ }
+ ]
+ }
+]