SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 151200 of 1878 papers

TitleStatusHype
A High-Quality Text-Rich Image Instruction Tuning Dataset via Hybrid Instruction GenerationCode0
Beyond Human Data: Aligning Multimodal Large Language Models by Iterative Self-EvolutionCode0
Toward Robust Hyper-Detailed Image Captioning: A Multiagent Approach and Dual Evaluation Metrics for Factuality and Coverage0
Reframing Image Difference Captioning with BLIP2IDC and Synthetic AugmentationCode0
Dataset Augmentation by Mixing Visual Concepts0
Unveiling Uncertainty: A Deep Dive into Calibration and Performance of Multimodal Large Language ModelsCode0
Flowing from Words to Pixels: A Framework for Cross-Modality Evolution0
Descriptive Caption Enhancement with Visual Specialists for Multimodal PerceptionCode0
G-VEval: A Versatile Metric for Evaluating Image and Video Captions Using GPT-4oCode1
JoVALE: Detecting Human Actions in Video Using Audiovisual and Language ContextsCode0
Typhoon 2: A Family of Open Text and Multimodal Thai Large Language ModelsCode1
Maybe you are looking for CroQS: Cross-modal Query Suggestion for Text-to-Image Retrieval0
MedMax: Mixed-Modal Instruction Tuning for Training Biomedical AssistantsCode1
PunchBench: Benchmarking MLLMs in Multimodal Punchline Comprehension0
UnMA-CapSumT: Unified and Multi-Head Attention-driven Caption Summarization Transformer0
Overview of TREC 2024 Medical Video Question Answering (MedVidQA) Track0
From Simple to Professional: A Combinatorial Controllable Image Captioning AgentCode0
Optimizing Vision-Language Interactions Through Decoder-Only Models0
Automated Image Captioning with CNNs and TransformersCode0
Vision-Language Models Represent Darker-Skinned Black Individuals as More Homogeneous than Lighter-Skinned Black Individuals0
Benchmarking Large Vision-Language Models via Directed Scene Graph for Comprehensive Image CaptioningCode1
Seeing Syntax: Uncovering Syntactic Learning Limitations in Vision-Language Models0
How Vision-Language Tasks Benefit from Large Pre-trained Models: A Survey0
3D Spatial Understanding in MLLMs: Disambiguation and Evaluation0
Exploring Multi-Grained Concept Annotations for Multimodal Large Language ModelsCode0
HMGIE: Hierarchical and Multi-Grained Inconsistency Evaluation for Vision-Language Data Cleansing0
Automated Medical Report Generation for ECG Data: Bridging Medical Text and Signal Processing with Deep LearningCode0
Florence-VL: Enhancing Vision-Language Models with Generative Vision Encoder and Depth-Breadth FusionCode3
Personalizing Multimodal Large Language Models for Image Captioning: An Experimental Analysis0
Remote Sensing Temporal Vision-Language Models: A Comprehensive SurveyCode3
Progress-Aware Video Frame Captioning0
CEGI: Measuring the trade-off between efficiency and carbon emissions for SLMs and VLMs0
DIR: Retrieval-Augmented Image Captioning with Comprehensive Understanding0
Improving Multimodal LLMs Ability In Geometry Problem Solving, Reasoning, And Multistep Scoring0
Sparse Attention Vectors: Generative Multimodal Model Features Are Discriminative Vision-Language Classifiers0
OPCap:Object-aware Prompting Captioning0
Active Data Curation Effectively Distills Large-Scale Multimodal Models0
Efficient Multi-modal Large Language Models via Visual Token Grouping0
LaB-RAG: Label Boosted Retrieval Augmented Generation for Radiology Report GenerationCode1
Debiasing Classifiers by Amplifying Bias with Latent Diffusion and Large Language Models0
Chain of Attack: On the Robustness of Vision-Language Models Against Transfer-Based Adversarial Attacks0
FINECAPTION: Compositional Image Captioning Focusing on Wherever You Want at Any Granularity0
FG-CXR: A Radiologist-Aligned Gaze Dataset for Enhancing Interpretability in Chest X-Ray Report GenerationCode1
Uterine Ultrasound Image Captioning Using Deep Learning Techniques0
LMM-driven Semantic Image-Text Coding for Ultra Low-bitrate Learned Image CompressionCode1
Mitigating Perception Bias: A Training-Free Approach to Enhance LMM for Image Quality Assessment0
AI Flow at the Network Edge0
The Power of Many: Multi-Agent Multimodal Models for Cultural Image CaptioningCode0
Learn from Downstream and Be Yourself in Multimodal Large Language Model Fine-TuningCode0
MolParser: End-to-end Visual Recognition of Molecule Structures in the Wild0
Show:102550
← PrevPage 4 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1IBM Research AICIDEr80.67Unverified
2CASIA_IVACIDEr79.15Unverified
3feixiangCIDEr77.31Unverified
4wocaoCIDEr77.21Unverified
5lamiwab172CIDEr75.93Unverified
6RUC_AIM3CIDEr73.52Unverified
7funasCIDEr73.51Unverified
8SRC-B_VCLabCIDEr73.47Unverified
9spartaCIDEr73.41Unverified
10x-vizCIDEr73.26Unverified
#ModelMetricClaimedVerifiedStatus
1VALORCIDER152.5Unverified
2VASTCIDER149Unverified
3Virtex (ResNet-101)CIDER94Unverified
4KOSMOS-1 (1.6B) (zero-shot)CIDER84.7Unverified
5BLIP-FuseCapCLIPScore78.5Unverified
6mPLUGBLEU-446.5Unverified
7OFABLEU-444.9Unverified
8GITBLEU-444.1Unverified
9BLIP-2 ViT-G OPT 2.7B (zero-shot)BLEU-443.7Unverified
10BLIP-2 ViT-G OPT 6.7B (zero-shot)BLEU-443.5Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr149.1Unverified
2GIT2, Single ModelCIDEr124.18Unverified
3GIT, Single ModelCIDEr122.4Unverified
4PaLICIDEr121.09Unverified
5CoCa - Google BrainCIDEr117.9Unverified
6Microsoft Cognitive Services teamCIDEr112.82Unverified
7Single ModelCIDEr108.98Unverified
8GRIT (zero-shot, no VL pretraining, no CBS)CIDEr105.9Unverified
9FudanFVLCIDEr104.9Unverified
10FudanWYZCIDEr104.25Unverified
#ModelMetricClaimedVerifiedStatus
1GIT2, Single ModelCIDEr125.51Unverified
2PaLICIDEr124.35Unverified
3GIT, Single ModelCIDEr123.92Unverified
4CoCa - Google BrainCIDEr120.73Unverified
5Microsoft Cognitive Services teamCIDEr115.54Unverified
6Single ModelCIDEr110.76Unverified
7FudanFVLCIDEr109.33Unverified
8FudanWYZCIDEr108.04Unverified
9IEDA-LABCIDEr100.15Unverified
10firetheholeCIDEr99.51Unverified
#ModelMetricClaimedVerifiedStatus
1PaLICIDEr126.67Unverified
2GIT2, Single ModelCIDEr122.27Unverified
3GIT, Single ModelCIDEr122.04Unverified
4CoCa - Google BrainCIDEr121.69Unverified
5Microsoft Cognitive Services teamCIDEr110.14Unverified
6Single ModelCIDEr109.49Unverified
7FudanFVLCIDEr106.55Unverified
8FudanWYZCIDEr103.75Unverified
9HumanCIDEr91.62Unverified
10firetheholeCIDEr88.54Unverified