@@ -25,7 +25,7 @@ RUN_HUGE_TESTS=false
25
25
if [ " ${1:- } " = " huge" ]; then
26
26
RUN_HUGE_TESTS=true
27
27
RUN_BIG_TESTS=true
28
- echo " Include BIG models..."
28
+ echo " Include BIG and HUGE models..."
29
29
fi
30
30
31
31
# ##############
93
93
# this will run both the big and huge models
94
94
# huge models are > 32B parameters
95
95
if [ " $RUN_HUGE_TESTS " = true ]; then
96
- add_test " llama-mtmd-cli " " ggml-org/Qwen2.5-VL-72B-Instruct-GGUF:Q4_K_M"
97
- add_test " llama-mtmd-cli " " ggml-org/Llama-4-Scout-17B-16E-Instruct-GGUF:IQ1_S"
96
+ add_test_vision " ggml-org/Qwen2.5-VL-72B-Instruct-GGUF:Q4_K_M"
97
+ add_test_vision " ggml-org/Llama-4-Scout-17B-16E-Instruct-GGUF:IQ1_S"
98
98
fi
99
99
100
100
# these models always give the wrong answer, not sure why
101
- # add_test "llama-mtmd-cli" "ggml-org/SmolVLM-Instruct-GGUF:Q4_K_M"
102
- # add_test "llama-mtmd-cli" "ggml-org/SmolVLM-256M-Instruct-GGUF:Q8_0"
103
- # add_test "llama-mtmd-cli" "ggml-org/SmolVLM2-256M-Video-Instruct-GGUF:Q8_0"
101
+ # add_test_vision "ggml-org/SmolVLM-Instruct-GGUF:Q4_K_M"
102
+ # add_test_vision "ggml-org/SmolVLM-256M-Instruct-GGUF:Q8_0"
103
+ # add_test_vision "ggml-org/SmolVLM2-256M-Video-Instruct-GGUF:Q8_0"
104
104
105
105
# this model has broken chat template, not usable
106
- # add_test "llama-mtmd-cli" "cmp-nct/Yi-VL-6B-GGUF:Q5_K"
107
- # add_test "llama-mtmd-cli" "guinmoon/MobileVLM-3B-GGUF:Q4_K_M" "deepseek"
106
+ # add_test_vision "cmp-nct/Yi-VL-6B-GGUF:Q5_K"
107
+ # add_test_vision "guinmoon/MobileVLM-3B-GGUF:Q4_K_M" "deepseek"
108
108
109
109
# ##############
110
110
0 commit comments