SOTAVerified

Video Visual Relation Detection

Video Visual Relation Detection (VidVRD) aims to detect instances of visual relations of interest in a video, where a visual relation instance is represented by a relation triplet with the trajectories of the subject and object. As compared to still images, videos provide a more natural set of features for detecting visual relations, such as the dynamic relations like “A-follow-B” and “A-towards-B”, and temporally changing relations like “A-chase-B” followed by “A-hold-B”. Yet, VidVRD is technically more challenging than ImgVRD due to the difficulties in accurate object tracking and diverse relation appearances in the video domain.

Source: ImageNet-VidVRD Video Visual Relation Dataset

Papers

Showing 115 of 15 papers

TitleStatusHype
VrdONE: One-stage Video Visual Relation DetectionCode1
SportsHHI: A Dataset for Human-Human Interaction Detection in Sports VideosCode1
Compositional Prompt Tuning with Motion Cues for Open-vocabulary Video Relation DetectionCode1
Video Relation Detection via Tracklet based Visual TransformerCode1
Spatial-Temporal Transformer for Dynamic Scene Graph GenerationCode1
What and When to Look?: Temporal Span Proposal Network for Video Relation DetectionCode1
LIGHTEN: Learning Interactions with Graph and Hierarchical TEmporal Networks for HOI in videosCode1
OpenVidVRD: Open-Vocabulary Video Visual Relation Detection via Prompt-Driven Semantic Space Alignment0
In Defense of Clip-based Video Relation Detection0
Meta Spatio-Temporal Debiasing for Video Scene Graph Generation0
VRDFormer: End-to-End Video Visual Relation Detection With Transformers0
Social Fabric: Tubelet Compositions for Video Relation DetectionCode0
Video Relation Detection with Trajectory-aware Multi-modal Features0
Beyond Short-Term Snippet: Video Relation Detection With Spatio-Temporal Global Context0
Video Relationship Reasoning using Gated Spatio-Temporal Energy GraphCode0
Show:102550

No leaderboard results yet.