SOTAVerified

Pose Manipulation with Identity Preservation

2020-04-20International Journal of Computers Communications & Control 2020Unverified0· sign in to hype

Andrei-Timotei Ardelean, Lucian Mircea Sasu

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

This paper describes a new model which generates images in novel poses e.g. by altering face expression and orientation, from just a few instances of a human subject. Unlike previous approaches which require large datasets of a specific person for training, our approach may start from a scarce set of images, even from a single image. To this end, we introduce Character Adaptive Identity Normalization GAN (CainGAN) which uses spatial characteristic features extracted by an embedder and combined across source images. The identity information is propagated throughout the network by applying conditional normalization. After extensive adversarial training, CainGAN receives figures of faces from a certain individual and produces new ones while preserving the person's identity. Experimental results show that the quality of generated images scales with the size of the input set used during inference. Furthermore, quantitative measurements indicate that CainGAN performs better compared to other methods when training data is limited.

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
VoxCeleb2 - 1-shot learningCainGANFID35Unverified
VoxCeleb2 - 8-shot learningCainGANFID24.9Unverified

Reproductions