SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 101150 of 1878 papers

TitleStatusHype
ReVision: A Dataset and Baseline VLM for Privacy-Preserving Task-Oriented Visual Instruction Rewriting0
What Is a Good Caption? A Comprehensive Visual Caption Benchmark for Evaluating Both Correctness and Thoroughness0
InsightVision: A Comprehensive, Multi-Level Chinese-based Benchmark for Evaluating Implicit Visual Semantics in Large Vision Language Models0
A Chain-of-Thought Subspace Meta-Learning for Few-shot Image Captioning with Large Vision and Language Models0
Pretrained Image-Text Models are Secretly Video CaptionersCode0
GroundCap: A Visually Grounded Image Captioning Dataset0
TPCap: Unlocking Zero-Shot Image Captioning with Trigger-Augmented and Multi-Modal Purification Modules0
VisCon-100K: Leveraging Contextual Web Data for Fine-tuning Vision Language Models0
GAIA: A Global, Multi-modal, Multi-scale Vision-Language Dataset for Remote Sensing Image AnalysisCode1
FE-LWS: Refined Image-Text Representations via Decoder Stacking and Fused Encodings for Remote Sensing Image Captioning0
Vision-Language Models for Edge Networks: A Comprehensive Survey0
Evaluation of Multilingual Image Captioning: How far can we get with CLIP models?Code0
Generative Distribution Prediction: A Unified Approach to Multimodal Learning0
Temporal Working Memory: Query-Guided Segment Refinement for Enhanced Multimodal UnderstandingCode3
Éclair -- Extracting Content and Layout with Integrated Reading Order for Documents0
Efficient Few-Shot Continual Learning in Vision-Language Models0
TexLiDAR: Automated Text Understanding for Panoramic LiDAR DataCode0
Exploring Spatial Language Grounding Through Referring Expressions0
COCONut-PanCap: Joint Panoptic Segmentation and Grounded Captions for Fine-Grained Understanding and Generation0
Robust-LLaVA: On the Effectiveness of Large-Scale Robust Image Encoders for Multi-modal Large Language ModelsCode1
MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding0
Large Vision-Language Models for Knowledge-Grounded Data Annotation of MemesCode0
An Ensemble Model with Attention Based Mechanism for Image Captioning0
PAINT: Paying Attention to INformed Tokens to Mitigate Hallucination in Large Vision-Language ModelCode1
Text-driven Adaptation of Foundation Models for Few-shot Surgical Workflow AnalysisCode0
LAVCap: LLM-based Audio-Visual Captioning using Optimal TransportCode1
Double Visual Defense: Adversarial Pre-training and Instruction Tuning for Improving Vision-Language Model Robustness0
VCRScore: Image captioning metric based on V\&L Transformers, CLIP, and precision-recall0
RadAlign: Advancing Radiology Report Generation with Vision-Language Concept AlignmentCode1
GeoPix: Multi-Modal Large Language Model for Pixel-level Image Understanding in Remote Sensing0
Valley2: Exploring Multimodal Models with Scalable Vision-Language DesignCode3
Improving Image Captioning by Mimicking Human Reformulation Feedback at Inference-time0
Evaluating Image Caption via Cycle-consistent Text-to-Image Generation0
Generalizing from SIMPLE to HARD Visual Reasoning: Can We Mitigate Modality Imbalance in VLMs?Code1
Decoding fMRI Data into Captions using Prefix Language ModelingCode0
MoColl: Agent-Based Specific and General Model Collaboration for Image Captioning0
Patch Matters: Training-free Fine-grained Image Caption Enhancement via Local Perception0
AdaDARE-gamma: Balancing Stability and Plasticity in Multi-modal LLMs through Efficient Adaptation0
Diffusion Bridge: Leveraging Diffusion Model to Reduce the Modality Gap Between Text and Vision for Zero-Shot Image CaptioningCode1
Semantic and Expressive Variations in Image Captions Across Languages0
Variance-Based Membership Inference Attacks Against Large-Scale Image Captioning Models0
Flowing from Words to Pixels: A Noise-Free Framework for Cross-Modality Evolution0
Unleashing Text-to-Image Diffusion Prior for Zero-Shot Image Captioning0
Enhanced Multimodal RAG-LLM for Accurate Visual Question Answering0
ErgoChat: a Visual Query System for the Ergonomic Risk Assessment of Construction Workers0
ViPCap: Retrieval Text-Based Visual Prompts for Lightweight Image CaptioningCode0
EvalMuse-40K: A Reliable and Fine-Grained Benchmark with Comprehensive Human Annotations for Text-to-Image Generation Model EvaluationCode2
Survey of Large Multimodal Model Datasets, Application Categories and Taxonomy0
GCS-M3VLT: Guided Context Self-Attention based Multi-modal Medical Vision Language Transformer for Retinal Image Captioning0
SilVar: Speech Driven Multimodal Model for Reasoning Visual Question Answering and Object LocalizationCode0
Show:102550
← PrevPage 3 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IBM Research AICIDEr80.67Unverified
2CASIA_IVACIDEr79.15Unverified
3feixiangCIDEr77.31Unverified
4wocaoCIDEr77.21Unverified
5lamiwab172CIDEr75.93Unverified
6RUC_AIM3CIDEr73.52Unverified
7funasCIDEr73.51Unverified
8SRC-B_VCLabCIDEr73.47Unverified
9spartaCIDEr73.41Unverified
10x-vizCIDEr73.26Unverified
#ModelMetricClaimedVerifiedStatus
1VALORCIDER152.5Unverified
2VASTCIDER149Unverified
3Virtex (ResNet-101)CIDER94Unverified
4KOSMOS-1 (1.6B) (zero-shot)CIDER84.7Unverified
5BLIP-FuseCapCLIPScore78.5Unverified
6mPLUGBLEU-446.5Unverified
7OFABLEU-444.9Unverified
8GITBLEU-444.1Unverified
9BLIP-2 ViT-G OPT 2.7B (zero-shot)BLEU-443.7Unverified
10BLIP-2 ViT-G OPT 6.7B (zero-shot)BLEU-443.5Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr149.1Unverified
2GIT2, Single ModelCIDEr124.18Unverified
3GIT, Single ModelCIDEr122.4Unverified
4PaLICIDEr121.09Unverified
5CoCa - Google BrainCIDEr117.9Unverified
6Microsoft Cognitive Services teamCIDEr112.82Unverified
7Single ModelCIDEr108.98Unverified
8GRIT (zero-shot, no VL pretraining, no CBS)CIDEr105.9Unverified
9FudanFVLCIDEr104.9Unverified
10FudanWYZCIDEr104.25Unverified
#ModelMetricClaimedVerifiedStatus
1GIT2, Single ModelCIDEr125.51Unverified
2PaLICIDEr124.35Unverified
3GIT, Single ModelCIDEr123.92Unverified
4CoCa - Google BrainCIDEr120.73Unverified
5Microsoft Cognitive Services teamCIDEr115.54Unverified
6Single ModelCIDEr110.76Unverified
7FudanFVLCIDEr109.33Unverified
8FudanWYZCIDEr108.04Unverified
9IEDA-LABCIDEr100.15Unverified
10firetheholeCIDEr99.51Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr126.67Unverified
2GIT2, Single ModelCIDEr122.27Unverified
3GIT, Single ModelCIDEr122.04Unverified
4CoCa - Google BrainCIDEr121.69Unverified
5Microsoft Cognitive Services teamCIDEr110.14Unverified
6Single ModelCIDEr109.49Unverified
7FudanFVLCIDEr106.55Unverified
8FudanWYZCIDEr103.75Unverified
9HumanCIDEr91.62Unverified
10firetheholeCIDEr88.54Unverified