SOTAVerified

Video Understanding

A crucial task of Video Understanding is to recognise and localise (in space and time) different actions or events appearing in the video.

Source: Action Detection from a Robot-Car Perspective

Papers

Showing 826850 of 1149 papers

TitleStatusHype
LLaVA-MLB: Mitigating and Leveraging Attention Bias for Training-Free Video LLMs0
LLaVA-Octopus: Unlocking Instruction-Driven Adaptive Projector Fusion for Video Understanding0
LLAVIDAL: A Large LAnguage VIsion Model for Daily Activities of Living0
LLM4Brain: Training a Large Language Model for Brain Video Understanding0
LLMs Meet Long Video: Advancing Long Video Question Answering with An Interactive Visual Adapter in LLMs0
Localizing Events in Videos with Multimodal Queries0
Localizing Unseen Activities in Video via Image Query0
Logic-in-Frames: Dynamic Keyframe Search via Visual Semantic-Logical Verification for Long Video Understanding0
Long Activity Video Understanding using Functional Object-Oriented Network0
LongCaptioning: Unlocking the Power of Long Caption Generation in Large Multimodal Models0
Long-Short Temporal Contrastive Learning of Video Transformers0
LongVILA: Scaling Long-Context Visual Language Models for Long Videos0
LongViTU: Instruction Tuning for Long-Form Video Understanding0
Long-VMNet: Accelerating Long-Form Video Understanding via Fixed Memory0
Look Every Frame All at Once: Video-Ma^2mba for Efficient Long-form Video Understanding with Multi-Axis Gradient Checkpointing0
Low-Fidelity End-to-End Video Encoder Pre-training for Temporal Action Localization0
LVAgent: Long Video Understanding by Multi-Round Dynamical Collaboration of MLLM Agents0
LV-XAttn: Distributed Cross-Attention for Long Visual Inputs in Multimodal Large Language Models0
M^33D: Learning 3D priors using Multi-Modal Masked Autoencoders for 2D image and video understanding0
M^3Net: Multi-view Encoding, Matching, and Fusion for Few-shot Fine-grained Action Recognition0
MaCP: Minimal yet Mighty Adaptation via Hierarchical Cosine Projection0
Making Every Frame Matter: Continuous Video Understanding for Large Models via Adaptive State Modeling0
MAMBA4D: Efficient Long-Sequence Point Cloud Video Understanding with Disentangled Spatial-Temporal State Space Models0
MambaMia: A State-Space-Model-Based Compression for Efficient Video Understanding in Large Multimodal Models0
MASH-VLM: Mitigating Action-Scene Hallucination in Video-LLMs through Disentangled Spatial-Temporal Representations0
Show:102550
← PrevPage 34 of 46Next →

No leaderboard results yet.