Evaluating Theory of Mind in Question Answering
2018-08-28EMNLP 2018Code Available1· sign in to hype
Aida Nematzadeh, Kaylee Burns, Erin Grant, Alison Gopnik, Thomas L. Griffiths
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/kayburns/tom-qa-datasetOfficialIn papernone★ 0
- github.com/GoodAI/goodai-ltm-benchmarknone★ 84
- github.com/facebookresearch/tominone★ 0
Abstract
We propose a new dataset for evaluating question answering models with respect to their capacity to reason about beliefs. Our tasks are inspired by theory-of-mind experiments that examine whether children are able to reason about the beliefs of others, in particular when those beliefs differ from reality. We evaluate a number of recent neural models with memory augmentation. We find that all fail on our tasks, which require keeping track of inconsistent states of the world; moreover, the models' accuracy decreases notably when random sentences are introduced to the tasks at test.