Video Grounding
Video grounding is the task of linking spoken language descriptions to specific video segments. In video grounding, the model is given a video and a natural language description, such as a sentence or a caption, and its goal is to identify the specific segment of the video that corresponds to the description. This can involve tasks such as localizing the objects or actions mentioned in the description within the video, or associating a specific time interval with the description.
Papers
Showing 1–10 of 114 papers
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | InternVideo2-6B | R@1,IoU=0.7 | 56.45 | — | Unverified |
| 2 | InternVideo2-1B | R@1,IoU=0.7 | 54.45 | — | Unverified |
| 3 | LLMEPET | R@1,IoU=0.7 | 49.94 | — | Unverified |
| 4 | QD-DETR | R@1,IoU=0.7 | 44.98 | — | Unverified |
| 5 | DiffusionVMR | R@1,IoU=0.7 | 44.49 | — | Unverified |
| 6 | UMT | R@1,IoU=0.7 | 41.18 | — | Unverified |
| 7 | Moment-DETR | R@1,IoU=0.7 | 33.02 | — | Unverified |
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | DeCafNet | R@1,IoU=0.1 | 13.25 | — | Unverified |
| 2 | DenoiseLoc | R@1,IoU=0.1 | 11.59 | — | Unverified |