SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 751800 of 1878 papers

TitleStatusHype
GPTs Are Multilingual Annotators for Sequence Generation TasksCode0
Exploring Visual Culture Awareness in GPT-4V: A Comprehensive Probing0
CIC: A Framework for Culturally-Aware Image Captioning0
Examining Gender and Racial Bias in Large Vision-Language Models Using a Novel Dataset of Parallel ImagesCode0
Image captioning for Brazilian Portuguese using GRIT model0
Text or Image? What is More Important in Cross-Domain Generalization Capabilities of Hate Meme Detection Models?0
PICS: Pipeline for Image Captioning and Search0
SCO-VIST: Social Interaction Commonsense Knowledge-based Visual Storytelling0
Good at captioning, bad at counting: Benchmarking GPT-4V on Earth observation dataCode0
COCO is "ALL'' You Need for Visual Instruction Fine-tuning0
KTVIC: A Vietnamese Image Captioning Dataset on the Life Domain0
Jewelry Recognition via Encoder-Decoder Models0
What Else Would I Like? A User Simulator using Alternatives for Improved Evaluation of Fashion Conversational Recommendation Systems0
Let's Go Shopping (LGS) -- Web-Scale Image-Text Dataset for Visual Concept Understanding0
MAMI: Multi-Attentional Mutual-Information for Long Sequence Neuron Captioning0
Hyperparameter-Free Approach for Faster Minimum Bayes Risk DecodingCode0
Object-oriented backdoor attack against image captioning0
SyCoCa: Symmetrizing Contrastive Captioners with Attentive Masking for Multimodal Alignment0
Social Media Ready Caption Generation for Brands0
Cycle-Consistency Learning for Captioning and Grounding0
LLM4VG: Large Language Models Evaluation for Video Grounding0
p-Laplacian Adaptation for Generative Pre-trained Vision-Language ModelsCode0
Dietary Assessment with Multimodal ChatGPT: A Systematic Analysis0
Improving Cross-modal Alignment with Synthetic Pairs for Text-only Image Captioning0
Synocene, Beyond the Anthropocene: De-Anthropocentralising Human-Nature-AI Interaction0
Filter & Align: Leveraging Human Knowledge to Curate Image-Text Data0
Unifying Text, Tables, and Images for Multimodal Question AnsweringCode0
PixLore: A Dataset-driven Approach to Rich Image CaptioningCode0
Lyrics: Boosting Fine-grained Language-Vision Alignment and Comprehension via Semantic-aware Visual Objects0
User-Aware Prefix-Tuning is a Good Learner for Personalized Image Captioning0
On the Robustness of Large Multimodal Models Against Image Adversarial Attacks0
Towards More Unified In-context Visual Understanding0
CLAMP: Contrastive LAnguage Model Prompt-tuning0
Automatic Report Generation for Histopathology images using pre-trained Vision Transformers and BERTCode0
Video Summarization: Towards Entity-Aware CaptionsCode0
Enhancing Image Captioning with Neural Models0
Omni-SMoLA: Boosting Generalist Multimodal Models with Soft Mixture of Low-rank Experts0
InstructSeq: Unifying Vision Tasks with Instruction-conditioned Multi-modal Sequence GenerationCode0
A natural language processing-based approach: mapping human perception by understanding deep semantic features in street view images0
MobileCLIP: Fast Image-Text Models through Multi-Modal Reinforced Training0
EVCap: Retrieval-Augmented Image Captioning with External Visual-Name Memory for Open-World Comprehension0
DECap: Towards Generalized Explicit Caption Editing via Diffusion Mechanism0
Violet: A Vision-Language Model for Arabic Image Captioning with Gemini Decoder0
Improving Image Captioning via Predicting Structured Concepts0
Holistic Evaluation of GPT-4V for Biomedical Imaging0
How to Bridge the Gap between Modalities: Survey on Multimodal Large Language Model0
Zero-shot Translation of Attention Patterns in VQA Models to Natural LanguageCode0
DeepPatent2: A Large-Scale Benchmarking Corpus for Technical Drawing UnderstandingCode0
JaSPICE: Automatic Evaluation Metric Using Predicate-Argument Structures for Image Captioning ModelsCode0
Visual Analytics for Efficient Image Exploration and User-Guided Image Captioning0
Show:102550
← PrevPage 16 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IBM Research AICIDEr80.67Unverified
2CASIA_IVACIDEr79.15Unverified
3feixiangCIDEr77.31Unverified
4wocaoCIDEr77.21Unverified
5lamiwab172CIDEr75.93Unverified
6RUC_AIM3CIDEr73.52Unverified
7funasCIDEr73.51Unverified
8SRC-B_VCLabCIDEr73.47Unverified
9spartaCIDEr73.41Unverified
10x-vizCIDEr73.26Unverified
#ModelMetricClaimedVerifiedStatus
1VALORCIDER152.5Unverified
2VASTCIDER149Unverified
3Virtex (ResNet-101)CIDER94Unverified
4KOSMOS-1 (1.6B) (zero-shot)CIDER84.7Unverified
5BLIP-FuseCapCLIPScore78.5Unverified
6mPLUGBLEU-446.5Unverified
7OFABLEU-444.9Unverified
8GITBLEU-444.1Unverified
9BLIP-2 ViT-G OPT 2.7B (zero-shot)BLEU-443.7Unverified
10BLIP-2 ViT-G OPT 6.7B (zero-shot)BLEU-443.5Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr149.1Unverified
2GIT2, Single ModelCIDEr124.18Unverified
3GIT, Single ModelCIDEr122.4Unverified
4PaLICIDEr121.09Unverified
5CoCa - Google BrainCIDEr117.9Unverified
6Microsoft Cognitive Services teamCIDEr112.82Unverified
7Single ModelCIDEr108.98Unverified
8GRIT (zero-shot, no VL pretraining, no CBS)CIDEr105.9Unverified
9FudanFVLCIDEr104.9Unverified
10FudanWYZCIDEr104.25Unverified
#ModelMetricClaimedVerifiedStatus
1GIT2, Single ModelCIDEr125.51Unverified
2PaLICIDEr124.35Unverified
3GIT, Single ModelCIDEr123.92Unverified
4CoCa - Google BrainCIDEr120.73Unverified
5Microsoft Cognitive Services teamCIDEr115.54Unverified
6Single ModelCIDEr110.76Unverified
7FudanFVLCIDEr109.33Unverified
8FudanWYZCIDEr108.04Unverified
9IEDA-LABCIDEr100.15Unverified
10firetheholeCIDEr99.51Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr126.67Unverified
2GIT2, Single ModelCIDEr122.27Unverified
3GIT, Single ModelCIDEr122.04Unverified
4CoCa - Google BrainCIDEr121.69Unverified
5Microsoft Cognitive Services teamCIDEr110.14Unverified
6Single ModelCIDEr109.49Unverified
7FudanFVLCIDEr106.55Unverified
8FudanWYZCIDEr103.75Unverified
9HumanCIDEr91.62Unverified
10firetheholeCIDEr88.54Unverified