RefineCap: Concept-Aware Refinement for Image Captioning
2021-09-08Unverified0· sign in to hype
Yekun Chai, Shuo Jin, Junliang Xing
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Automatically translating images to texts involves image scene understanding and language modeling. In this paper, we propose a novel model, termed RefineCap, that refines the output vocabulary of the language decoder using decoder-guided visual semantics, and implicitly learns the mapping between visual tag words and images. The proposed Visual-Concept Refinement method can allow the generator to attend to semantic details in the image, thereby generating more semantically descriptive captions. Our model achieves superior performance on the MS-COCO dataset in comparison with previous visual-concept based models.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| COCO Captions | RefineCap (w/ REINFORCE) | BLEU-4 | 37.8 | — | Unverified |