Researchers propose a new approach for enhancing reasoning capabilities in Multimodal Large Language Models (MLLMs).Effective cold start initialization is identified as crucial for improving MLLM reasoning, even before applying multimodal reinforcement learning.Standard GRPO used in multimodal reinforcement learning faces issues like gradient stagnation, impacting training stability and performance.A staged training approach called ReVisual-R1 is introduced, achieving a new state-of-the-art performance on various challenging benchmarks.