SOTAVerified

Image Captioning

Image Captioning is the task of describing the content of an image in words. This task lies at the intersection of computer vision and natural language processing. Most image captioning systems use an encoder-decoder framework, where an input image is encoded into an intermediate representation of the information in the image, and then decoded into a descriptive text sequence. The most popular benchmarks are nocaps and COCO, and models are typically evaluated according to a BLEU or CIDER metric.

( Image credit: Reflective Decoding Network for Image Captioning, ICCV'19)

Papers

Showing 110 of 1878 papers

Show:102550
← PrevPage 1 of 188Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1BLIP-2 ViT-G FlanT5 XL (zero-shot)CIDEr123.7Unverified
2BLIP-2 ViT-G OPT 6.7B (zero-shot)CIDEr123.7Unverified
3BLIP-2 ViT-G OPT 2.7B (zero-shot)CIDEr123Unverified
4LEMON_largeCIDEr116.9Unverified
5BLIP_ViT-LCIDEr114.9Unverified
6SimVLMCIDEr113.7Unverified
7BLIP_CapFilt-LCIDEr111.8Unverified
8LEMON_baseCIDEr107.7Unverified
9OmniVLCIDEr104.6Unverified
10VinVLCIDEr103.1Unverified