SOTAVerified

Image-text matching

Image-Text Matching is a subtask within Cross-Modal Retrieval (CMR) that involves establishing associations between images and corresponding textual descriptions. The goal is to retrieve an image given a textual query or, conversely, retrieve a textual description given an image query. This task is challenging due to the heterogeneity gap between image and text data representations. Image-text matching is used in applications such as content-based image search, visual question answering, and multimodal summarization.

Assessing Brittleness of Image-Text Retrieval Benchmarks from Vision-Language Models Perspective

Papers

Showing 176188 of 188 papers

TitleStatusHype
ImageBERT: Cross-modal Pre-training with Large-scale Weak-supervised Image-Text Data0
Learning fragment self-attention embeddings for image-text matchingCode0
UNITER: Learning UNiversal Image-TExt Representations0
Learning Visual Relation Priors for Image-Text Matching and Image Captioning with Neural Scene Graph Generators0
Unicoder-VL: A Universal Encoder for Vision and Language by Cross-modal Pre-training0
Matching Images and Text with Multi-modal Tensor Fusion and Re-rankingCode0
Knowledge Aware Semantic Concept Expansion for Image-Text Matching0
Position Focused Attention Network for Image-Text MatchingCode0
ParNet: Position-aware Aggregated Relation Network for Image-Text matching0
Deep Cross-Modal Projection Learning for Image-Text MatchingCode0
Cross-modal Subspace Learning for Fine-grained Sketch-based Image Retrieval0
Learning Two-Branch Neural Networks for Image-Text Matching TasksCode0
Dual Attention Networks for Multimodal Reasoning and MatchingCode0
Show:102550
← PrevPage 8 of 8Next →

No leaderboard results yet.