SOTAVerified

Video Understanding

A crucial task of Video Understanding is to recognise and localise (in space and time) different actions or events appearing in the video.

Source: Action Detection from a Robot-Car Perspective

Papers

Showing 926950 of 1149 papers

TitleStatusHype
Learning without Prejudice: Avoiding Bias in Webly-Supervised Action Recognition0
Less than Few: Self-Shot Video Instance Segmentation0
Leveraging Foundation Models for Multimodal Graph-Based Action Recognition0
Leveraging Local Temporal Information for Multimodal Scene Classification0
LIGAR: Lightweight General-purpose Action Recognition0
LiveVLM: Efficient Online Video Understanding via Streaming-Oriented KV Cache and Retrieval0
LiVLR: A Lightweight Visual-Linguistic Reasoning Framework for Video Question Answering0
LLaVA-MLB: Mitigating and Leveraging Attention Bias for Training-Free Video LLMs0
LLaVA-Octopus: Unlocking Instruction-Driven Adaptive Projector Fusion for Video Understanding0
LLAVIDAL: A Large LAnguage VIsion Model for Daily Activities of Living0
LLM4Brain: Training a Large Language Model for Brain Video Understanding0
LLMs Meet Long Video: Advancing Long Video Question Answering with An Interactive Visual Adapter in LLMs0
Localizing Events in Videos with Multimodal Queries0
Localizing Unseen Activities in Video via Image Query0
Logic-in-Frames: Dynamic Keyframe Search via Visual Semantic-Logical Verification for Long Video Understanding0
Long Activity Video Understanding using Functional Object-Oriented Network0
LongCaptioning: Unlocking the Power of Long Caption Generation in Large Multimodal Models0
Long-Short Temporal Contrastive Learning of Video Transformers0
LongVILA: Scaling Long-Context Visual Language Models for Long Videos0
LongViTU: Instruction Tuning for Long-Form Video Understanding0
Long-VMNet: Accelerating Long-Form Video Understanding via Fixed Memory0
Look Every Frame All at Once: Video-Ma^2mba for Efficient Long-form Video Understanding with Multi-Axis Gradient Checkpointing0
Low-Fidelity End-to-End Video Encoder Pre-training for Temporal Action Localization0
LVAgent: Long Video Understanding by Multi-Round Dynamical Collaboration of MLLM Agents0
LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models0
Show:102550
← PrevPage 38 of 46Next →

No leaderboard results yet.