SOTAVerified

Visual Grounding

Visual Grounding (VG) aims to locate the most relevant object or region in an image, based on a natural language query. The query can be a phrase, a sentence, or even a multi-round dialogue. There are three main challenges in VG:

  • What is the main focus in a query?
  • How to understand an image?
  • How to locate an object?

Papers

Showing 301325 of 571 papers

TitleStatusHype
Improved Visual Grounding through Self-Consistent Explanations0
Improving Visually Grounded Sentence Representations with Self-Attention0
Individuation in Neural Models with and without Visual Grounding0
Intent3D: 3D Object Detection in RGB-D Scans Based on Human Intention0
Interactive Reinforcement Learning for Object Grounding via Self-Talking0
Interactive Visual Grounding of Referring Expressions for Human-Robot Interaction0
Interpretable Visual Question Answering by Visual Grounding from Attention Supervision Mining0
Interpretable Visual Question Answering via Reasoning Supervision0
INVIGORATE: Interactive Visual Grounding and Grasping in Clutter0
I Speak and You Find: Robust 3D Visual Grounding with Noisy and Ambiguous Speech Inputs0
Joint Top-Down and Bottom-Up Frameworks for 3D Visual Grounding0
Knowledge Supports Visual Language Grounding: A Case Study on Colour Terms0
Language-Guided 3D Object Detection in Point Cloud for Autonomous Driving0
Language learning using Speech to Image retrieval0
LanguageRefer: Spatial-Language Model for 3D Visual Grounding0
LCV2: An Efficient Pretraining-Free Framework for Grounded Visual Question Answering0
Learning from Synthetic Data for Visual Grounding0
Visually Consistent Hierarchical Image Classification0
Learning Language Structures through Grounding0
Learning to Compose and Reason with Language Tree Structures for Visual Grounding0
Learning to Ground VLMs without Forgetting0
Learning Unsupervised Visual Grounding Through Semantic Self-Supervision0
Learning Visual Grounding from Generative Vision and Language Model0
Learning with Difference Attention for Visually Grounded Self-supervised Representations0
Less is More: Generating Grounded Navigation Instructions from Landmarks0
Show:102550
← PrevPage 13 of 23Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Florence-2-large-ftAccuracy (%)95.3Unverified
2mPLUG-2Accuracy (%)92.8Unverified
3X2-VLM (large)Accuracy (%)92.1Unverified
4XFM (base)Accuracy (%)90.4Unverified
5X2-VLM (base)Accuracy (%)90.3Unverified
6X-VLM (base)Accuracy (%)89Unverified
7HYDRAIoU61.7Unverified
8HYDRAIoU61.1Unverified
#ModelMetricClaimedVerifiedStatus
1Florence-2-large-ftAccuracy (%)92Unverified
2mPLUG-2Accuracy (%)86.05Unverified
3X2-VLM (large)Accuracy (%)81.8Unverified
4XFM (base)Accuracy (%)79.8Unverified
5X2-VLM (base)Accuracy (%)78.4Unverified
6X-VLM (base)Accuracy (%)76.91Unverified
#ModelMetricClaimedVerifiedStatus
1Florence-2-large-ftAccuracy (%)93.4Unverified
2mPLUG-2Accuracy (%)90.33Unverified
3X2-VLM (large)Accuracy (%)87.6Unverified
4XFM (base)Accuracy (%)86.1Unverified
5X2-VLM (base)Accuracy (%)85.2Unverified
6X-VLM (base)Accuracy (%)84.51Unverified