SOTAVerified

Multimodal Differential Network for Visual Question Generation

2018-08-12EMNLP 2018Code Available0· sign in to hype

Badri N. Patro, Sandeep Kumar, Vinod K. Kurmi, Vinay P. Namboodiri

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Generating natural questions from an image is a semantic task that requires using visual and language modality to learn multimodal representations. Images can have multiple visual and language contexts that are relevant for generating questions namely places, captions, and tags. In this paper, we propose the use of exemplars for obtaining the relevant context. We obtain this by using a Multimodal Differential Network to produce natural and engaging questions. The generated questions show a remarkable similarity to the natural questions as validated by a human study. Further, we observe that the proposed approach substantially improves over state-of-the-art benchmarks on the quantitative metrics (BLEU, METEOR, ROUGE, and CIDEr).

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
COCO Visual Question Answering (VQA) real images 1.0 open endedMDNBLEU-165.1Unverified
Visual Question GenerationMDNBLEU-136Unverified

Reproductions