SOTAVerified

Learning Human-Object Interactions by Graph Parsing Neural Networks

2018-08-23ECCV 2018Code Available0· sign in to hype

Siyuan Qi, Wenguan Wang, Baoxiong Jia, Jianbing Shen, Song-Chun Zhu

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper addresses the task of detecting and recognizing human-object interactions (HOI) in images and videos. We introduce the Graph Parsing Neural Network (GPNN), a framework that incorporates structural knowledge while being differentiable end-to-end. For a given scene, GPNN infers a parse graph that includes i) the HOI graph structure represented by an adjacency matrix, and ii) the node labels. Within a message passing inference framework, GPNN iteratively computes the adjacency matrices and node labels. We extensively evaluate our model on three HOI detection benchmarks on images and videos: HICO-DET, V-COCO, and CAD-120 datasets. Our approach significantly outperforms state-of-art methods, verifying that GPNN is scalable to large datasets and applies to spatial-temporal settings. The code is available at https://github.com/SiyuanQi/gpnn.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
HICO-DETGPNNmAP13.11Unverified
V-COCOGPNNAP(S1)44Unverified

Reproductions