SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 51100 of 1878 papers

TitleStatusHype
VHM: Versatile and Honest Vision Language Model for Remote Sensing Image AnalysisCode2
Can Language Beat Numerical Regression? Language-Based Multimodal Trajectory PredictionCode2
VL-ICL Bench: The Devil in the Details of Multimodal In-Context LearningCode2
Beyond Text: Frozen Large Language Models in Visual Signal ComprehensionCode2
MeaCap: Memory-Augmented Zero-shot Image CaptioningCode2
VTG-GPT: Tuning-Free Zero-Shot Video Temporal Grounding with GPTCode2
Learning Vision from Models Rivals Learning Vision from DataCode2
VCoder: Versatile Vision Encoders for Multimodal Large Language ModelsCode2
LLaMA-VID: An Image is Worth 2 Tokens in Large Language ModelsCode2
GLaMM: Pixel Grounding Large Multimodal ModelCode2
From CLIP to DINO: Visual Encoders Shout in Multi-modal Large Language ModelsCode2
MathVista: Evaluating Mathematical Reasoning of Foundation Models in Visual ContextsCode2
Fine-tuning Multimodal LLMs to Follow Zero-shot Demonstrative InstructionsCode2
JourneyDB: A Benchmark for Generative Image UnderstandingCode2
LLaVAR: Enhanced Visual Instruction Tuning for Text-Rich Image UnderstandingCode2
Shikra: Unleashing Multimodal LLM's Referential Dialogue MagicCode2
LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language ModelsCode2
Scalable 3D Captioning with Pretrained ModelsCode2
Contextual Object Detection with Multimodal Large Language ModelsCode2
VAST: A Vision-Audio-Subtitle-Text Omni-Modality Foundation Model and DatasetCode2
BiomedGPT: A Generalist Vision-Language Foundation Model for Diverse Biomedical TasksCode2
VALOR: Vision-Audio-Language Omni-Perception Pretraining Model and DatasetCode2
ChatGPT Asks, BLIP-2 Answers: Automatic Questioning Towards Enriched Visual DescriptionsCode2
Semantic-Conditional Diffusion Networks for Image CaptioningCode2
X^2-VLM: All-In-One Pre-trained Model For Vision-Language TasksCode2
Text-Only Training for Image Captioning using Noise-Injected CLIPCode2
PoseScript: Linking 3D Human Poses and Natural LanguageCode2
Visual Language Maps for Robot NavigationCode2
Open-Vocabulary Semantic Segmentation with Mask-adapted CLIPCode2
Pix2Struct: Screenshot Parsing as Pretraining for Visual Language UnderstandingCode2
Comprehending and Ordering Semantics for Image CaptioningCode2
Uni-Perceiver-MoE: Learning Sparse Generalist Models with Conditional MoEsCode2
GIT: A Generative Image-to-text Transformer for Vision and LanguageCode2
Fine-grained Image Captioning with CLIP RewardCode2
Language Models Can See: Plugging Visual Controls in Text GenerationCode2
ClipCap: CLIP Prefix for Image CaptioningCode2
VinVL: Revisiting Visual Representations in Vision-Language ModelsCode2
Controlling Length in Image CaptioningCode2
Oscar: Object-Semantics Aligned Pre-training for Vision-Language TasksCode2
A Better Variant of Self-Critical Sequence TrainingCode2
Unified Vision-Language Pre-Training for Image Captioning and VQACode2
ViCrit: A Verifiable Reinforcement Learning Proxy Task for Visual Perception in VLMsCode1
DiscoVLA: Discrepancy Reduction in Vision, Language, and Alignment for Parameter-Efficient Video-Text RetrievalCode1
Puzzled by Puzzles: When Vision-Language Models Can't Take a HintCode1
SATORI-R1: Incentivizing Multimodal Reasoning with Spatial Grounding and Verifiable RewardsCode1
SilVar-Med: A Speech-Driven Visual Language Model for Explainable Abnormality Detection in Medical ImagingCode1
A Survey on Efficient Vision-Language ModelsCode1
Mind the Gap: Benchmarking Spatial Reasoning in Vision-Language ModelsCode1
Image Captioning Evaluation in the Age of Multimodal LLMs: Challenges and Future PerspectivesCode1
Will Pre-Training Ever End? A First Step Toward Next-Generation Foundation MLLMs via Self-Improving Systematic CognitionCode1
Show:102550
← PrevPage 2 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IBM Research AICIDEr80.67Unverified
2CASIA_IVACIDEr79.15Unverified
3feixiangCIDEr77.31Unverified
4wocaoCIDEr77.21Unverified
5lamiwab172CIDEr75.93Unverified
6RUC_AIM3CIDEr73.52Unverified
7funasCIDEr73.51Unverified
8SRC-B_VCLabCIDEr73.47Unverified
9spartaCIDEr73.41Unverified
10x-vizCIDEr73.26Unverified
#ModelMetricClaimedVerifiedStatus
1VALORCIDER152.5Unverified
2VASTCIDER149Unverified
3Virtex (ResNet-101)CIDER94Unverified
4KOSMOS-1 (1.6B) (zero-shot)CIDER84.7Unverified
5BLIP-FuseCapCLIPScore78.5Unverified
6mPLUGBLEU-446.5Unverified
7OFABLEU-444.9Unverified
8GITBLEU-444.1Unverified
9BLIP-2 ViT-G OPT 2.7B (zero-shot)BLEU-443.7Unverified
10BLIP-2 ViT-G OPT 6.7B (zero-shot)BLEU-443.5Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr149.1Unverified
2GIT2, Single ModelCIDEr124.18Unverified
3GIT, Single ModelCIDEr122.4Unverified
4PaLICIDEr121.09Unverified
5CoCa - Google BrainCIDEr117.9Unverified
6Microsoft Cognitive Services teamCIDEr112.82Unverified
7Single ModelCIDEr108.98Unverified
8GRIT (zero-shot, no VL pretraining, no CBS)CIDEr105.9Unverified
9FudanFVLCIDEr104.9Unverified
10FudanWYZCIDEr104.25Unverified
#ModelMetricClaimedVerifiedStatus
1GIT2, Single ModelCIDEr125.51Unverified
2PaLICIDEr124.35Unverified
3GIT, Single ModelCIDEr123.92Unverified
4CoCa - Google BrainCIDEr120.73Unverified
5Microsoft Cognitive Services teamCIDEr115.54Unverified
6Single ModelCIDEr110.76Unverified
7FudanFVLCIDEr109.33Unverified
8FudanWYZCIDEr108.04Unverified
9IEDA-LABCIDEr100.15Unverified
10firetheholeCIDEr99.51Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr126.67Unverified
2GIT2, Single ModelCIDEr122.27Unverified
3GIT, Single ModelCIDEr122.04Unverified
4CoCa - Google BrainCIDEr121.69Unverified
5Microsoft Cognitive Services teamCIDEr110.14Unverified
6Single ModelCIDEr109.49Unverified
7FudanFVLCIDEr106.55Unverified
8FudanWYZCIDEr103.75Unverified
9HumanCIDEr91.62Unverified
10firetheholeCIDEr88.54Unverified