SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 451500 of 1878 papers

TitleStatusHype
Building Trustworthy Multimodal AI: A Review of Fairness, Transparency, and Ethics in Vision-Language Tasks0
AeroLite: Tag-Guided Lightweight Generation of Aerial Image Captions0
Metropolis-Hastings Captioning Game: Knowledge Fusion of Vision Language Models via Decentralized Bayesian Inference0
Embodied Image Captioning: Self-supervised Learning Agents for Spatially Coherent Image Descriptions0
AstroLLaVA: towards the unification of astronomical data and natural language0
Impact of Language Guidance: A Reproducibility Study0
How Can Objects Help Video-Language Understanding?0
RS-RAG: Bridging Remote Sensing Imagery and Comprehensive Knowledge with a Multi-Modal Dataset and Retrieval-Augmented Generation Model0
MORAL: A Multimodal Reinforcement Learning Framework for Decision Making in Autonomous Laboratories0
Group-based Distinctive Image Captioning with Memory Difference Encoding and Attention0
A Conformal Risk Control Framework for Granular Word Assessment and Uncertainty Calibration of CLIPScore Quality Estimates0
Context-Independent OCR with Multimodal LLMs: Effects of Image Resolution and Visual Complexity0
Semantic-Spatial Feature Fusion with Dynamic Graph Refinement for Remote Sensing Image Captioning0
JEEM: Vision-Language Understanding in Four Arabic Dialects0
Mitigating Low-Level Visual Hallucinations Requires Self-Awareness: Database, Model and Training Strategy0
Improved Alignment of Modalities in Large Vision Language Models0
Reverse Prompt: Cracking the Recipe Inside Text-to-Image Generation0
UniCrossAdapter: Multimodal Adaptation of CLIP for Radiology Report GenerationCode0
Natural Language Generation0
Disentangling Fine-Tuning from Pre-Training in Visual Captioning with Hybrid Markov LogicCode0
Unified Autoregressive Visual Generation and Understanding with Continuous Tokens0
GeoRSMLLM: A Multimodal Large Language Model for Vision-Language Tasks in Geoscience and Remote Sensing0
CapArena: Benchmarking and Analyzing Detailed Image Captioning in the LLM Era0
RONA: Pragmatically Diverse Image Captioning with Coherence RelationsCode0
Taxonomic Reasoning for Rare Arthropods: Combining Dense Image Captioning and RAG for Interpretable Classification0
Florenz: Scaling Laws for Systematic Generalization in Vision-Language Models0
Astrea: A MOE-based Visual Understanding Model with Progressive Alignment0
ComicsPAP: understanding comic strips by picking the correct panel0
Measuring directional bias amplification in image captions using predictability0
Improving cognitive diagnostics in pathology: a deep learning approach for augmenting perceptional understanding of histopathology images0
PerturboLLaVA: Reducing Multimodal Hallucinations with Perturbative Visual Training0
From Captions to Rewards (CAREVL): Leveraging Large Language Model Experts for Enhanced Reward Modeling in Large Vision-Language Models0
Treble Counterfactual VLMs: A Causal Approach to HallucinationCode0
A Benchmark for Multi-Lingual Vision-Language Learning in Remote Sensing Image CaptioningCode0
Group Relative Policy Optimization for Image CaptioningCode0
AC-Lite : A Lightweight Image Captioning Model for Low-Resource Assamese Language0
Exploring Causes and Mitigation of Hallucinations in Large Vision Language Models0
Are Large Language Models Good Data Preprocessors?0
Fine-Grained Video Captioning through Scene Graph Consolidation0
Good Representation, Better Explanation: Role of Convolutional Neural Networks in Transformer-Based Remote Sensing Image Captioning0
ReVision: A Dataset and Baseline VLM for Privacy-Preserving Task-Oriented Visual Instruction Rewriting0
A Chain-of-Thought Subspace Meta-Learning for Few-shot Image Captioning with Large Vision and Language Models0
InsightVision: A Comprehensive, Multi-Level Chinese-based Benchmark for Evaluating Implicit Visual Semantics in Large Vision Language Models0
GroundCap: A Visually Grounded Image Captioning Dataset0
Pretrained Image-Text Models are Secretly Video CaptionersCode0
What Is a Good Caption? A Comprehensive Visual Caption Benchmark for Evaluating Both Correctness and Thoroughness0
TPCap: Unlocking Zero-Shot Image Captioning with Trigger-Augmented and Multi-Modal Purification Modules0
VisCon-100K: Leveraging Contextual Web Data for Fine-tuning Vision Language Models0
FE-LWS: Refined Image-Text Representations via Decoder Stacking and Fused Encodings for Remote Sensing Image Captioning0
Vision-Language Models for Edge Networks: A Comprehensive Survey0
Show:102550
← PrevPage 10 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IBM Research AICIDEr80.67Unverified
2CASIA_IVACIDEr79.15Unverified
3feixiangCIDEr77.31Unverified
4wocaoCIDEr77.21Unverified
5lamiwab172CIDEr75.93Unverified
6RUC_AIM3CIDEr73.52Unverified
7funasCIDEr73.51Unverified
8SRC-B_VCLabCIDEr73.47Unverified
9spartaCIDEr73.41Unverified
10x-vizCIDEr73.26Unverified
#ModelMetricClaimedVerifiedStatus
1VALORCIDER152.5Unverified
2VASTCIDER149Unverified
3Virtex (ResNet-101)CIDER94Unverified
4KOSMOS-1 (1.6B) (zero-shot)CIDER84.7Unverified
5BLIP-FuseCapCLIPScore78.5Unverified
6mPLUGBLEU-446.5Unverified
7OFABLEU-444.9Unverified
8GITBLEU-444.1Unverified
9BLIP-2 ViT-G OPT 2.7B (zero-shot)BLEU-443.7Unverified
10BLIP-2 ViT-G OPT 6.7B (zero-shot)BLEU-443.5Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr149.1Unverified
2GIT2, Single ModelCIDEr124.18Unverified
3GIT, Single ModelCIDEr122.4Unverified
4PaLICIDEr121.09Unverified
5CoCa - Google BrainCIDEr117.9Unverified
6Microsoft Cognitive Services teamCIDEr112.82Unverified
7Single ModelCIDEr108.98Unverified
8GRIT (zero-shot, no VL pretraining, no CBS)CIDEr105.9Unverified
9FudanFVLCIDEr104.9Unverified
10FudanWYZCIDEr104.25Unverified
#ModelMetricClaimedVerifiedStatus
1GIT2, Single ModelCIDEr125.51Unverified
2PaLICIDEr124.35Unverified
3GIT, Single ModelCIDEr123.92Unverified
4CoCa - Google BrainCIDEr120.73Unverified
5Microsoft Cognitive Services teamCIDEr115.54Unverified
6Single ModelCIDEr110.76Unverified
7FudanFVLCIDEr109.33Unverified
8FudanWYZCIDEr108.04Unverified
9IEDA-LABCIDEr100.15Unverified
10firetheholeCIDEr99.51Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr126.67Unverified
2GIT2, Single ModelCIDEr122.27Unverified
3GIT, Single ModelCIDEr122.04Unverified
4CoCa - Google BrainCIDEr121.69Unverified
5Microsoft Cognitive Services teamCIDEr110.14Unverified
6Single ModelCIDEr109.49Unverified
7FudanFVLCIDEr106.55Unverified
8FudanWYZCIDEr103.75Unverified
9HumanCIDEr91.62Unverified
10firetheholeCIDEr88.54Unverified