SOTAVerified

Bridge Video and Text with Cascade Syntactic Structure

2018-08-01COLING 2018Unverified0· sign in to hype

Guolong Wang, Zheng Qin, Kaiping Xu, Kai Huang, Shuxiong Ye

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We present a video captioning approach that encodes features by progressively completing syntactic structure (LSTM-CSS). To construct basic syntactic structure (i.e., subject, predicate, and object), we use a Conditional Random Field to label semantic representations (i.e., motions, objects). We argue that in order to improve the comprehensiveness of the description, the local features within object regions can be used to generate complementary syntactic elements (e.g., attribute, adverbial). Inspired by redundancy of human receptors, we utilize a Region Proposal Network to focus on the object regions. To model the final temporal dynamics, Recurrent Neural Network with Path Embeddings is adopted. We demonstrate the effectiveness of LSTM-CSS on generating natural sentences: 42.3\% and 28.5\% in terms of BLEU@4 and METEOR. Superior performance when compared to state-of-the-art methods are reported on a large video description dataset (i.e., MSR-VTT-2016).

Tasks

Reproductions