< Explain other AI papers

Vision-R1: Incentivizing Reasoning Capability in Multimodal Large Language Models

Wenxuan Huang, Bohan Jia, Zijie Zhai, Shaosheng Cao, Zheyu Ye, Fei Zhao, Yao Hu, Shaohui Lin

2025-03-11

Vision-R1: Incentivizing Reasoning Capability in Multimodal Large
  Language Models

Summary

This paper talks about Vision-R1, an AI tool that teaches computers to solve image-based math problems by learning step-by-step reasoning like a student, using practice questions and feedback to improve.

What's the problem?

AI struggles with complex image-and-text problems (like math diagrams) because it either skips steps or overcomplicates answers when trained with basic methods, leading to mistakes.

What's the solution?

Vision-R1 uses a two-step method: first, it learns from a big set of example problems with clear answers, then it practices with feedback to focus on efficient reasoning without overthinking.

Why it matters?

This helps AI tutors and assistants solve real-world problems (like homework help or science diagrams) more accurately and explain their steps clearly, making them more trustworthy.

Abstract

DeepSeek-R1-Zero has successfully demonstrated the emergence of reasoning capabilities in LLMs purely through Reinforcement Learning (RL). Inspired by this breakthrough, we explore how RL can be utilized to enhance the reasoning capability of MLLMs. However, direct training with RL struggles to activate complex reasoning capabilities such as questioning and reflection in MLLMs, due to the absence of substantial high-quality multimodal reasoning data. To address this issue, we propose the reasoning MLLM, Vision-R1, to improve multimodal reasoning capability. Specifically, we first construct a high-quality multimodal CoT dataset without human annotations by leveraging an existing MLLM and DeepSeek-R1 through modality bridging and data filtering to obtain a 200K multimodal CoT dataset, Vision-R1-cold dataset. It serves as cold-start initialization data for Vision-R1. To mitigate the optimization challenges caused by overthinking after cold start, we propose Progressive Thinking Suppression Training (PTST) strategy and employ Group Relative Policy Optimization (GRPO) with the hard formatting result reward function to gradually refine the model's ability to learn correct and complex reasoning processes on a 10K multimodal math dataset. Comprehensive experiments show our model achieves an average improvement of sim6% across various multimodal math reasoning benchmarks. Vision-R1-7B achieves a 73.5% accuracy on the widely used MathVista benchmark, which is only 0.4% lower than the leading reasoning model, OpenAI O1. The datasets and code will be released in: https://github.com/Osilly/Vision-R1 .