SOTAVerified

Video Understanding

A crucial task of Video Understanding is to recognise and localise (in space and time) different actions or events appearing in the video.

Source: Action Detection from a Robot-Car Perspective

Papers

Showing 191200 of 1149 papers

TitleStatusHype
On the Limitations of Vision-Language Models in Understanding Image Transforms0
Measure Twice, Cut Once: Grasping Video Structures and Event Semantics with LLMs for Video Temporal Localization0
FaVChat: Unlocking Fine-Grained Facail Video Understanding with Multimodal Large Language Models0
Everything Can Be Described in Words: A Simple Unified Multi-Modal Framework with Semantic and Temporal Alignment0
VideoScan: Enabling Efficient Streaming Video Understanding via Frame-level Semantic Carriers0
VLog: Video-Language Models by Generative Retrieval of Narration VocabularyCode4
Exo2Ego: Exocentric Knowledge Guided MLLM for Egocentric Video Understanding0
Generative Frame Sampler for Long Video Understanding0
Memory-enhanced Retrieval Augmentation for Long Video Understanding0
QuoTA: Query-oriented Token Assignment via CoT Query Decouple for Long Video ComprehensionCode2
Show:102550
← PrevPage 20 of 115Next →

No leaderboard results yet.