From f33fd766c9cd0ec7fc0aace8e9f32f9733c3eed7 Mon Sep 17 00:00:00 2001 From: n1ck-guo Date: Tue, 14 Oct 2025 03:13:43 -0400 Subject: [PATCH] fix vlm gguf ut Signed-off-by: n1ck-guo --- test/test_cpu/test_gguf_format.py | 1 - 1 file changed, 1 deletion(-) diff --git a/test/test_cpu/test_gguf_format.py b/test/test_cpu/test_gguf_format.py index 96b489ffe..d71920b39 100644 --- a/test/test_cpu/test_gguf_format.py +++ b/test/test_cpu/test_gguf_format.py @@ -329,7 +329,6 @@ def test_vlm_gguf(self): autoround.quantize_and_save(output_dir=quantized_model_path, format="gguf:q4_0") self.assertTrue("mmproj-model.gguf" in os.listdir("./saved")) for file_name in os.listdir(quantized_model_path): - file_name = os.listdir(quantized_model_path)[0] file_size = os.path.getsize(os.path.join(quantized_model_path, file_name)) / 1024**2 if file_name == "mmproj-model.gguf": self.assertAlmostEqual(file_size, 2535, delta=1.0)