SOTAVerified

Actor and Action Video Segmentation from a Sentence

2018-03-20CVPR 2018Code Available1· sign in to hype

Kirill Gavrilyuk, Amir Ghodrati, Zhenyang Li, Cees G. M. Snoek

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

This paper strives for pixel-level segmentation of actors and their actions in video content. Different from existing works, which all learn to segment from a fixed vocabulary of actor and action pairs, we infer the segmentation from a natural language input sentence. This allows to distinguish between fine-grained actors in the same super-category, identify actor and action instances, and segment pairs that are outside of the actor and action vocabulary. We propose a fully-convolutional model for pixel-level actor and action segmentation using an encoder-decoder architecture optimized for video. To show the potential of actor and action video segmentation from a sentence, we extend two popular actor and action datasets with more than 7,500 natural language descriptions. Experiments demonstrate the quality of the sentence-guided segmentations, the generalization ability of our model, and its advantage for traditional actor and action segmentation compared to the state-of-the-art.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
A2D SentencesGavriluyk el al. (Optical flow)AP0.22Unverified
A2D SentencesGavriluyk el al.AP0.2Unverified
J-HMDBGavrilyuk et al. (Optical flow)AP0.27Unverified
J-HMDBGavrilyuk et al.AP0.23Unverified

Reproductions