IQBench: How "Smart'' Are Vision-Language Models? A Study with Human IQ Tests
Tan-Hanh Pham, Phu-Vinh Nguyen, Dang The Hung, Bui Trong Duong, Vu Nguyen Thanh, Chris Ngo, Tri Quang Truong, Truong-Son Hy
2025-05-29
Summary
This paper talks about IQBench, a new way to test how 'smart' AI models that work with both pictures and words really are, by seeing how well they do on human IQ test questions that involve visual patterns.
What's the problem?
The problem is that while AI models can sometimes give the right answer, it's hard to know if they're actually reasoning like a human or just guessing. Most tests only check if the final answer is correct, without looking at how the AI got there or if it can explain its thinking.
What's the solution?
The researchers created IQBench, a benchmark that checks not just if the AI gets the right answer, but also how well it recognizes patterns and explains its reasoning. This gives a better idea of how the AI thinks and whether it's really understanding the questions like a person would.
Why it matters?
This is important because it helps us figure out how close AI is to human-level thinking, and what still needs to be improved. It can also lead to smarter, more trustworthy AI that can explain its answers, which is useful for education, science, and everyday problem-solving.
Abstract
IQBench evaluates the reasoning capabilities of Vision-Language Models on visual IQ tests, focusing on pattern recognition and explanation quality rather than final accuracy alone.