SOTAVerified

Tag Disentangled Generative Adversarial Networks for Object ImageRe-rendering

2017-02-19International Joint Conference on Artificial Intelligence 2017Unverified0· sign in to hype

Chaoyue Wang, Chaohui Wang, Chang Xu, DaCheng Tao

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

In this paper, we propose a principled Tag Dis-entangled Generative Adversarial Networks (TD-GAN) for re-rendering new images for the object of interest from a single image of it by specifying multiple scene properties (such as viewpoint, illumination, expression,etc.). The whole framework consists of a disentangling network, a generative network, a tag mapping net, and a discriminative network, which are trained jointly based on a given set of images that are complete/partially tagged(i.e., supervised/semi-supervised setting). Given an input image, the disentangling network extracts dis-entangled and interpretable representations, which are then used to generate images by the generative network. In order to boost the quality of disentangled representations, the tag mapping net is integrated to explore the consistency between the image and its tags. Furthermore, the discriminative network is introduced to implement the adversarial training strategy for generating more realistic images. Experiments on two challenging datasets demonstrate the state-of-the-art performance of the proposed framework in the problem of interest.

Tasks

Reproductions