Document Type

Conference Proceeding

Publication Title

Advances in Neural Information Processing Systems

Abstract

This paper proposes a human parsing based texture transfer model via cross-view consistency learning to generate the texture of 3D human body from a single image. We use the semantic parsing of human body as input for providing both the shape and pose information to reduce the appearance variation of human image and preserve the spatial distribution of semantic parts. Meanwhile, in order to improve the prediction for textures of invisible parts, we explicitly enforce the consistency across different views of the same subject by exchanging the textures predicted by two views to render images during training. The perceptual loss and total variation regularization are optimized to maximize the similarity between rendered and input images, which does not necessitate extra 3D texture supervision. Experimental results on pedestrian images and fashion photos demonstrate that our method can produce higher quality textures with convincing details than other texture generation methods. Code is available at https://github.com/zhaofang0627/HPBTT.

Publication Date

12-2020

Keywords

Image enhancement, Rendering (computer graphics), Semantics, Textures

Comments

Access available at NeurIPS 2020 proceedings online

Share

COinS