SOTAVerified

Video Grounding

Video grounding is the task of linking spoken language descriptions to specific video segments. In video grounding, the model is given a video and a natural language description, such as a sentence or a caption, and its goal is to identify the specific segment of the video that corresponds to the description. This can involve tasks such as localizing the objects or actions mentioned in the description within the video, or associating a specific time interval with the description.

Papers

Showing 101110 of 114 papers

TitleStatusHype
Semi-Supervised Video Paragraph Grounding With Contrastive Encoder0
Seq2Time: Sequential Knowledge Transfer for Video LLM Temporal Grounding0
SimBase: A Simple Baseline for Temporal Video Grounding0
Simplify Implant Depth Prediction as Video Grounding: A Texture Perceive Implant Depth Prediction Network0
SpaceVLLM: Endowing Multimodal Large Language Model with Spatio-Temporal Video Grounding Capability0
SpikeMba: Multi-Modal Spiking Saliency Mamba for Temporal Video Grounding0
STPro: Spatial and Temporal Progressive Learning for Weakly Supervised Spatio-Temporal Grounding0
STVGBert: A Visual-Linguistic Transformer Based Framework for Spatio-Temporal Video Grounding0
STVGFormer: Spatio-Temporal Video Grounding with Static-Dynamic Cross-Modal Understanding0
Support-Set Based Cross-Supervision for Video Grounding0
Show:102550
← PrevPage 11 of 12Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1InternVideo2-6BR@1,IoU=0.756.45Unverified
2InternVideo2-1BR@1,IoU=0.754.45Unverified
3LLMEPETR@1,IoU=0.749.94Unverified
4QD-DETRR@1,IoU=0.744.98Unverified
5DiffusionVMRR@1,IoU=0.744.49Unverified
6UMTR@1,IoU=0.741.18Unverified
7Moment-DETRR@1,IoU=0.733.02Unverified
#ModelMetricClaimedVerifiedStatus
1DeCafNetR@1,IoU=0.113.25Unverified
2DenoiseLocR@1,IoU=0.111.59Unverified