SOTAVerified

MVBench

Papers

Showing 119 of 19 papers

TitleStatusHype
CogVLM2: Visual Language Models for Image and Video UnderstandingCode9
Video-R1: Reinforcing Video Reasoning in MLLMsCode4
PLLaVA : Parameter-free LLaVA Extension from Images to Videos for Video Dense CaptioningCode4
Flash-VStream: Efficient Real-Time Understanding for Long Video StreamsCode3
VideoChat-R1: Enhancing Spatio-Temporal Perception via Reinforcement Fine-TuningCode3
Lyra: An Efficient and Speech-Centric Framework for Omni-CognitionCode3
VideoGPT+: Integrating Image and Video Encoders for Enhanced Video UnderstandingCode3
Mobile-VideoGPT: Fast and Accurate Video Understanding Language ModelCode2
LLaVAction: evaluating and training multi-modal large language models for action recognitionCode2
Video-CCAM: Enhancing Video-Language Understanding with Causal Cross-Attention Masks for Short and Long VideosCode2
ST-LLM: Large Language Models Are Effective Temporal LearnersCode2
MVBench: A Comprehensive Multi-modal Video Understanding BenchmarkCode2
VideoHallu: Evaluating and Mitigating Multi-modal Hallucinations on Synthetic Video UnderstandingCode1
TS-LLaVA: Constructing Visual Tokens through Thumbnail-and-Sampling for Training-Free Video Large Language ModelsCode1
GAM-Agent: Game-Theoretic and Uncertainty-Aware Collaboration for Complex Visual Reasoning0
VideoPASTA: 7K Preference Pairs That Matter for Video-LLM Alignment0
VideoSAVi: Self-Aligned Video Language Models without Human Supervision0
Enhancing Temporal Modeling of Video LLMs via Time GatingCode0
VideoLLaMB: Long-context Video Understanding with Recurrent Memory Bridges0
Show:102550

No leaderboard results yet.