SOTAVerified

Factor Graph Attention

2019-04-11CVPR 2019Code Available0· sign in to hype

Idan Schwartz, Seunghak Yu, Tamir Hazan, Alexander Schwing

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Dialog is an effective way to exchange information, but subtle details and nuances are extremely important. While significant progress has paved a path to address visual dialog with algorithms, details and nuances remain a challenge. Attention mechanisms have demonstrated compelling results to extract details in visual question answering and also provide a convincing framework for visual dialog due to their interpretability and effectiveness. However, the many data utilities that accompany visual dialog challenge existing attention techniques. We address this issue and develop a general attention mechanism for visual dialog which operates on any number of data utilities. To this end, we design a factor graph based attention mechanism which combines any number of utility representations. We illustrate the applicability of the proposed approach on the challenging and recently introduced VisDial datasets, outperforming recent state-of-the-art methods by 1.1% for VisDial0.9 and by 2% for VisDial1.0 on MRR. Our ensemble model improved the MRR score on VisDial1.0 by more than 6%.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
VisDial v0.9 val9xFGA (VGG)MRR68.92Unverified
Visual Dialog v1.0 test-std5xFGA (F-RCNNx101)NDCG (x 100)57.2Unverified

Reproductions