SOTAVerified

Video Understanding

A crucial task of Video Understanding is to recognise and localise (in space and time) different actions or events appearing in the video.

Source: Action Detection from a Robot-Car Perspective

Papers

Showing 281290 of 1149 papers

TitleStatusHype
Language-Guided Audio-Visual Learning for Long-Term Sports AssessmentCode1
VEU-Bench: Towards Comprehensive Understanding of Video Editing0
Weakly Supervised Temporal Action Localization via Dual-Prior Collaborative Learning Guided by Multimodal Large Language Models0
OV-HHIR: Open Vocabulary Human Interaction Recognition Using Cross-modal Integration of Large Language Models0
CaReBench: A Fine-Grained Benchmark for Video Captioning and Retrieval0
Embodied VideoAgent: Persistent Memory from Egocentric Videos and Embodied Sensors Enables Dynamic Scene Understanding0
VideoRefer Suite: Advancing Spatial-Temporal Object Understanding with Video LLMCode3
Online Video Understanding: OVBench and VideoChat-OnlineCode2
FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Visual Language ModelsCode2
Detection-Fusion for Knowledge Graph Extraction from VideosCode0
Show:102550
← PrevPage 29 of 115Next →

No leaderboard results yet.