SOTAVerified

Image-text matching

Image-Text Matching is a subtask within Cross-Modal Retrieval (CMR) that involves establishing associations between images and corresponding textual descriptions. The goal is to retrieve an image given a textual query or, conversely, retrieve a textual description given an image query. This task is challenging due to the heterogeneity gap between image and text data representations. Image-text matching is used in applications such as content-based image search, visual question answering, and multimodal summarization.

Assessing Brittleness of Image-Text Retrieval Benchmarks from Vision-Language Models Perspective

Papers

Showing 176188 of 188 papers

TitleStatusHype
Visual Semantic Reasoning for Image-Text MatchingCode1
VL-BERT: Pre-training of Generic Visual-Linguistic RepresentationsCode1
Unicoder-VL: A Universal Encoder for Vision and Language by Cross-modal Pre-training0
Matching Images and Text with Multi-modal Tensor Fusion and Re-rankingCode0
Knowledge Aware Semantic Concept Expansion for Image-Text Matching0
Position Focused Attention Network for Image-Text MatchingCode0
ParNet: Position-aware Aggregated Relation Network for Image-Text matching0
Deep Cross-Modal Projection Learning for Image-Text MatchingCode0
Stacked Cross Attention for Image-Text MatchingCode1
AttnGAN: Fine-Grained Text to Image Generation with Attentional Generative Adversarial NetworksCode1
Cross-modal Subspace Learning for Fine-grained Sketch-based Image Retrieval0
Learning Two-Branch Neural Networks for Image-Text Matching TasksCode0
Dual Attention Networks for Multimodal Reasoning and MatchingCode0
Show:102550
← PrevPage 8 of 8Next →

No leaderboard results yet.