SOTAVerified

Video Grounding

Video grounding is the task of linking spoken language descriptions to specific video segments. In video grounding, the model is given a video and a natural language description, such as a sentence or a caption, and its goal is to identify the specific segment of the video that corresponds to the description. This can involve tasks such as localizing the objects or actions mentioned in the description within the video, or associating a specific time interval with the description.

Papers

Showing 5175 of 114 papers

TitleStatusHype
Seq2Time: Sequential Knowledge Transfer for Video LLM Temporal Grounding0
SimBase: A Simple Baseline for Temporal Video Grounding0
SynopGround: A Large-Scale Dataset for Multi-Paragraph Video Grounding from TV Dramas and Synopses0
Multi-sentence Video Grounding for Long Video Generation0
Described Spatial-Temporal Video Detection0
AutoTVG: A New Vision-language Pre-training Paradigm for Temporal Video Grounding0
Simplify Implant Depth Prediction as Video Grounding: A Texture Perceive Implant Depth Prediction Network0
Artemis: Towards Referential Understanding in Complex VideosCode0
Video-of-Thought: Step-by-Step Video Reasoning from Perception to Cognition0
SpikeMba: Multi-Modal Spiking Saliency Mamba for Temporal Video Grounding0
Unified Static and Dynamic Network: Efficient Temporal Filtering for Video GroundingCode0
VideoGrounding-DINO: Towards Open-Vocabulary Spatio-Temporal Video Grounding0
Video-GroundingDINO: Towards Open-Vocabulary Spatio-Temporal Video Grounding0
Multi-Modal Domain Adaptation Across Video Scenes for Temporal Video Grounding0
LLM4VG: Large Language Models Evaluation for Video Grounding0
Cross-modal Contrastive Learning with Asymmetric Co-attention Network for Video Moment RetrievalCode0
EtC: Temporal Boundary Expand then Clarify for Weakly Supervised Video Grounding with Multimodal Large Language Model0
Exploring Iterative Refinement with Diffusion Models for Video GroundingCode0
Dual-Path Temporal Map Optimization for Make-up Temporal Video GroundingCode0
DiffusionVMR: Diffusion Model for Joint Video Moment Retrieval and Highlight Detection0
ViGT: Proposal-free Video Grounding with Learnable Token in Transformer0
G2L: Semantically Aligned and Uniform Video Grounding via Geodesic and Game Theory0
No-frills Temporal Video Grounding: Multi-Scale Neighboring Attention and Zoom-in Boundary Detection0
Dense Video Object Captioning from Disjoint SupervisionCode0
Boundary-Denoising for Video Activity LocalizationCode0
Show:102550
← PrevPage 3 of 5Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1InternVideo2-6BR@1,IoU=0.756.45Unverified
2InternVideo2-1BR@1,IoU=0.754.45Unverified
3LLMEPETR@1,IoU=0.749.94Unverified
4QD-DETRR@1,IoU=0.744.98Unverified
5DiffusionVMRR@1,IoU=0.744.49Unverified
6UMTR@1,IoU=0.741.18Unverified
7Moment-DETRR@1,IoU=0.733.02Unverified
#ModelMetricClaimedVerifiedStatus
1DeCafNetR@1,IoU=0.113.25Unverified
2DenoiseLocR@1,IoU=0.111.59Unverified