SOTAVerified

RefVOS: A Closer Look at Referring Expressions for Video Object Segmentation

2020-10-01Code Available1· sign in to hype

Miriam Bellver, Carles Ventura, Carina Silberer, Ioannis Kazakos, Jordi Torres, Xavier Giro-i-Nieto

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

The task of video object segmentation with referring expressions (language-guided VOS) is to, given a linguistic phrase and a video, generate binary masks for the object to which the phrase refers. Our work argues that existing benchmarks used for this task are mainly composed of trivial cases, in which referents can be identified with simple phrases. Our analysis relies on a new categorization of the phrases in the DAVIS-2017 and Actor-Action datasets into trivial and non-trivial REs, with the non-trivial REs annotated with seven RE semantic categories. We leverage this data to analyze the results of RefVOS, a novel neural network that obtains competitive results for the task of language-guided image segmentation and state of the art results for language-guided VOS. Our study indicates that the major challenges for the task are related to understanding motion and static actions.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
A2Dre testRefVosOverall IoU47.5Unverified
A2D SentencesRefVOSIoU overall0.6Unverified
A2D SentencesRefVOSIoU overall0.67Unverified
DAVIS 2017 (val)RefVOSJ&F 1st frame44.5Unverified
DAVIS 2017 (val)RefVOSJ&F 1st frame45.1Unverified
RefCOCO testARefVOS with BERT + MLM LossOverall IoU49.73Unverified
RefCOCO+ test BRefVOS with BERT + MLM lossOverall IoU36.17Unverified
RefCoCo valRefVOS with BERT + MLM lossOverall IoU59.45Unverified
RefCoCo valRefVOS with BERT + MLM lossOverall IoU44.71Unverified
RefCoCo valRefVOS with BERT Pre-trainOverall IoU58.65Unverified

Reproductions