SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 351400 of 1878 papers

TitleStatusHype
CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-ExpertsCode2
Using Machine Translation to Augment Multilingual Classification0
LLM as Dataset Analyst: Subpopulation Structure Discovery with Large Language ModelCode0
A Toolchain for Comprehensive Audio/Video Analysis Using Deep Learning Based Multimodal Approach (A use case of riot or violent context detection)0
Technical Report of NICE Challenge at CVPR 2024: Caption Re-ranking Evaluation Using Ensembled CLIP and Consensus ScoresCode0
Beyond Human Vision: The Role of Large Vision Language Models in Microscope Image Analysis0
What Makes for Good Image Captions?0
Compressed Image Captioning using CNN-based Encoder-Decoder Framework0
Semi-supervised Text-based Person Search0
Learning text-to-video retrieval from image captioning0
OmniSearchSage: Multi-Task Multi-Entity Embeddings for Pinterest SearchCode2
Lost in Space: Probing Fine-grained Spatial Understanding in Vision and Language ResamplersCode0
The Solution for the CVPR2024 NICE Image Captioning Challenge0
MM-PhyRLHF: Reinforcement Learning Framework for Multimodal Physics Question-Answering0
LaDiC: Are Diffusion Models Really Inferior to Autoregressive Counterparts for Image-to-Text Generation?Code1
ANCHOR: LLM-driven News Subject Conditioning for Text-to-Image SynthesisCode0
Bridging Vision and Language Spaces with Assignment PredictionCode0
On Speculative Decoding for Multimodal Large Language Models0
FLoRA: Enhancing Vision-Language Models with Parameter-Efficient Federated LearningCode0
Enhancing Visual Question Answering through Question-Driven Image Captions as PromptsCode1
View Selection for 3D Captioning via Diffusion RankingCode3
Panoptic Perception: A Novel Task and Fine-grained Dataset for Universal Remote Sensing Image Interpretation0
CoMat: Aligning Text-to-Image Diffusion Model with Image-to-Text Concept MatchingCode2
Would Deep Generative Models Amplify Bias in Future Models?0
Jump Self-attention: Capturing High-order Statistics in Transformers0
Harnessing the Power of Large Vision Language Models for Synthetic Image DetectionCode1
Disentangled Pre-training for Human-Object Interaction DetectionCode1
Bi-LORA: A Vision-Language Approach for Synthetic Image DetectionCode1
VLRM: Vision-Language Models act as Reward Models for Image Captioning0
Learning by Correction: Efficient Tuning Task for Zero-Shot Generative Vision-Language ReasoningCode0
LLaMA-Excitor: General Instruction Tuning via Indirect Feature Interaction0
VHM: Versatile and Honest Vision Language Model for Remote Sensing Image AnalysisCode2
A Review of Multi-Modal Large Language and Vision Models0
LocCa: Visual Pretraining with Location-aware CaptionersCode0
Text Data-Centric Image Captioning with Interactive Prompts0
Semantic Map-based Generation of Navigation InstructionsCode0
A Survey on Large Language Models from Concept to Implementation0
Can Language Beat Numerical Regression? Language-Based Multimodal Trajectory PredictionCode2
Automated Report Generation for Lung Cytological Images Using a CNN Vision Classifier and Multiple-Transformer Text Decoders: Preliminary Study0
Semi-Supervised Image Captioning Considering Wasserstein Graph Matching0
Visual Hallucination: Definition, Quantification, and Prescriptive Remediations0
The Solution for the ICCV 2023 1st Scientific Figure Captioning Challenge0
Image Captioning in news report scenario0
Cognitive resilience: Unraveling the proficiency of image-captioning models to interpret masked visual contentCode0
A Multimodal Approach for Cross-Domain Image Retrieval0
MyVLM: Personalizing VLMs for User-Specific Queries0
Inserting Faces inside Captions: Image Captioning with Attention Guided Merging0
Improved Baselines for Data-efficient Perceptual Augmentation of LLMs0
VL-ICL Bench: The Devil in the Details of Multimodal In-Context LearningCode2
Entity6K: A Large Open-Domain Evaluation Dataset for Real-World Entity Recognition0
Show:102550
← PrevPage 8 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IBM Research AICIDEr80.67Unverified
2CASIA_IVACIDEr79.15Unverified
3feixiangCIDEr77.31Unverified
4wocaoCIDEr77.21Unverified
5lamiwab172CIDEr75.93Unverified
6RUC_AIM3CIDEr73.52Unverified
7funasCIDEr73.51Unverified
8SRC-B_VCLabCIDEr73.47Unverified
9spartaCIDEr73.41Unverified
10x-vizCIDEr73.26Unverified
#ModelMetricClaimedVerifiedStatus
1VALORCIDER152.5Unverified
2VASTCIDER149Unverified
3Virtex (ResNet-101)CIDER94Unverified
4KOSMOS-1 (1.6B) (zero-shot)CIDER84.7Unverified
5BLIP-FuseCapCLIPScore78.5Unverified
6mPLUGBLEU-446.5Unverified
7OFABLEU-444.9Unverified
8GITBLEU-444.1Unverified
9BLIP-2 ViT-G OPT 2.7B (zero-shot)BLEU-443.7Unverified
10BLIP-2 ViT-G OPT 6.7B (zero-shot)BLEU-443.5Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr149.1Unverified
2GIT2, Single ModelCIDEr124.18Unverified
3GIT, Single ModelCIDEr122.4Unverified
4PaLICIDEr121.09Unverified
5CoCa - Google BrainCIDEr117.9Unverified
6Microsoft Cognitive Services teamCIDEr112.82Unverified
7Single ModelCIDEr108.98Unverified
8GRIT (zero-shot, no VL pretraining, no CBS)CIDEr105.9Unverified
9FudanFVLCIDEr104.9Unverified
10FudanWYZCIDEr104.25Unverified
#ModelMetricClaimedVerifiedStatus
1GIT2, Single ModelCIDEr125.51Unverified
2PaLICIDEr124.35Unverified
3GIT, Single ModelCIDEr123.92Unverified
4CoCa - Google BrainCIDEr120.73Unverified
5Microsoft Cognitive Services teamCIDEr115.54Unverified
6Single ModelCIDEr110.76Unverified
7FudanFVLCIDEr109.33Unverified
8FudanWYZCIDEr108.04Unverified
9IEDA-LABCIDEr100.15Unverified
10firetheholeCIDEr99.51Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr126.67Unverified
2GIT2, Single ModelCIDEr122.27Unverified
3GIT, Single ModelCIDEr122.04Unverified
4CoCa - Google BrainCIDEr121.69Unverified
5Microsoft Cognitive Services teamCIDEr110.14Unverified
6Single ModelCIDEr109.49Unverified
7FudanFVLCIDEr106.55Unverified
8FudanWYZCIDEr103.75Unverified
9HumanCIDEr91.62Unverified
10firetheholeCIDEr88.54Unverified