DiscoverDaily Paper CastVIDEOP2R: Video Understanding from Perception to Reasoning
VIDEOP2R: Video Understanding from Perception to Reasoning

VIDEOP2R: Video Understanding from Perception to Reasoning

Update: 2025-11-20
Share

Description

🤗 Upvotes: 70 | cs.CV, cs.AI, cs.LG



Authors:

Yifan Jiang, Yueying Wang, Rui Zhao, Toufiq Parag, Zhimin Chen, Zhenyu Liao, Jayakrishnan Unnikrishnan



Title:

VIDEOP2R: Video Understanding from Perception to Reasoning



Arxiv:

http://arxiv.org/abs/2511.11113v1



Abstract:

Reinforcement fine-tuning (RFT), a two-stage framework consisting of supervised fine-tuning (SFT) and reinforcement learning (RL) has shown promising results on improving reasoning ability of large language models (LLMs). Yet extending RFT to large video language models (LVLMs) remains challenging. We propose VideoP2R, a novel process-aware video RFT framework that enhances video reasoning by modeling perception and reasoning as distinct processes. In the SFT stage, we develop a three-step pipeline to generate VideoP2R-CoT-162K, a high-quality, process-aware chain-of-thought (CoT) dataset for perception and reasoning. In the RL stage, we introduce a novel process-aware group relative policy optimization (PA-GRPO) algorithm that supplies separate rewards for perception and reasoning. Extensive experiments show that VideoP2R achieves state-of-the-art (SotA) performance on six out of seven video reasoning and understanding benchmarks. Ablation studies further confirm the effectiveness of our process-aware modeling and PA-GRPO and demonstrate that model's perception output is information-sufficient for downstream reasoning.

Comments 
In Channel
loading
00:00
00:00
x

0.5x

0.8x

1.0x

1.25x

1.5x

2.0x

3.0x

Sleep Timer

Off

End of Episode

5 Minutes

10 Minutes

15 Minutes

30 Minutes

45 Minutes

60 Minutes

120 Minutes

VIDEOP2R: Video Understanding from Perception to Reasoning

VIDEOP2R: Video Understanding from Perception to Reasoning

Jingwen Liang, Gengyu Wang