SOTAVerified

Reasoning Visual Dialogs with Structural and Partial Observations

2019-04-11CVPR 2019Code Available0· sign in to hype

Zilong Zheng, Wenguan Wang, Siyuan Qi, Song-Chun Zhu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We propose a novel model to address the task of Visual Dialog which exhibits complex dialog structures. To obtain a reasonable answer based on the current question and the dialog history, the underlying semantic dependencies between dialog entities are essential. In this paper, we explicitly formalize this task as inference in a graphical model with partially observed nodes and unknown graph structures (relations in dialog). The given dialog entities are viewed as the observed nodes. The answer to a given question is represented by a node with missing value. We first introduce an Expectation Maximization algorithm to infer both the underlying dialog structures and the missing node values (desired answers). Based on this, we proceed to propose a differentiable graph neural network (GNN) solution that approximates this process. Experiment results on the VisDial and VisDial-Q datasets show that our model outperforms comparative methods. It is also observed that our method can infer the underlying dialog structure for better dialog reasoning.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
VisDial v0.9 valGNNMRR0.63Unverified
Visual Dialog v1.0 test-stdGNNNDCG (x 100)52.82Unverified

Reproductions