SOTAVerified

Video Grounding

Video grounding is the task of linking spoken language descriptions to specific video segments. In video grounding, the model is given a video and a natural language description, such as a sentence or a caption, and its goal is to identify the specific segment of the video that corresponds to the description. This can involve tasks such as localizing the objects or actions mentioned in the description within the video, or associating a specific time interval with the description.

Papers

Showing 7180 of 114 papers

TitleStatusHype
Language-free Training for Zero-shot Video Grounding0
Weakly-Supervised Temporal Article GroundingCode1
Graph2Vid: Flow graph to Video Grounding for Weakly-supervised Multi-Step Localization0
On the Effects of Video Grounding on Language Models0
Embracing Consistency: A One-Stage Approach for Spatio-Temporal Video GroundingCode1
Towards Parameter-Efficient Integration of Pre-Trained Language Models In Temporal Video GroundingCode0
CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal GroundingCode1
Video-Guided Curriculum Learning for Spoken Video GroundingCode0
Exploiting Feature Diversity for Make-up Temporal Video Grounding0
Team PKU-WICT-MIPL PIC Makeup Temporal Video Grounding Challenge 2022 Technical Report0
Show:102550
← PrevPage 8 of 12Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1InternVideo2-6BR@1,IoU=0.756.45Unverified
2InternVideo2-1BR@1,IoU=0.754.45Unverified
3LLMEPETR@1,IoU=0.749.94Unverified
4QD-DETRR@1,IoU=0.744.98Unverified
5DiffusionVMRR@1,IoU=0.744.49Unverified
6UMTR@1,IoU=0.741.18Unverified
7Moment-DETRR@1,IoU=0.733.02Unverified
#ModelMetricClaimedVerifiedStatus
1DeCafNetR@1,IoU=0.113.25Unverified
2DenoiseLocR@1,IoU=0.111.59Unverified