@@ -25,7 +25,7 @@ RUN_HUGE_TESTS=false
2525if [ " ${1:- } " = " huge" ]; then
2626 RUN_HUGE_TESTS=true
2727 RUN_BIG_TESTS=true
28- echo " Include BIG models..."
28+ echo " Include BIG and HUGE models..."
2929fi
3030
3131# ##############
9393# this will run both the big and huge models
9494# huge models are > 32B parameters
9595if [ " $RUN_HUGE_TESTS " = true ]; then
96- add_test " llama-mtmd-cli " " ggml-org/Qwen2.5-VL-72B-Instruct-GGUF:Q4_K_M"
97- add_test " llama-mtmd-cli " " ggml-org/Llama-4-Scout-17B-16E-Instruct-GGUF:IQ1_S"
96+ add_test_vision " ggml-org/Qwen2.5-VL-72B-Instruct-GGUF:Q4_K_M"
97+ add_test_vision " ggml-org/Llama-4-Scout-17B-16E-Instruct-GGUF:IQ1_S"
9898fi
9999
100100# these models always give the wrong answer, not sure why
101- # add_test "llama-mtmd-cli" "ggml-org/SmolVLM-Instruct-GGUF:Q4_K_M"
102- # add_test "llama-mtmd-cli" "ggml-org/SmolVLM-256M-Instruct-GGUF:Q8_0"
103- # add_test "llama-mtmd-cli" "ggml-org/SmolVLM2-256M-Video-Instruct-GGUF:Q8_0"
101+ # add_test_vision "ggml-org/SmolVLM-Instruct-GGUF:Q4_K_M"
102+ # add_test_vision "ggml-org/SmolVLM-256M-Instruct-GGUF:Q8_0"
103+ # add_test_vision "ggml-org/SmolVLM2-256M-Video-Instruct-GGUF:Q8_0"
104104
105105# this model has broken chat template, not usable
106- # add_test "llama-mtmd-cli" "cmp-nct/Yi-VL-6B-GGUF:Q5_K"
107- # add_test "llama-mtmd-cli" "guinmoon/MobileVLM-3B-GGUF:Q4_K_M" "deepseek"
106+ # add_test_vision "cmp-nct/Yi-VL-6B-GGUF:Q5_K"
107+ # add_test_vision "guinmoon/MobileVLM-3B-GGUF:Q4_K_M" "deepseek"
108108
109109# ##############
110110
0 commit comments