Document Type
Conference Proceeding
Publication Title
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
Abstract
Recent advances in 3D-aware generative models (3D-aware GANs) combined with Neural Radiance Fields (NeRF) have achieved impressive results. However no prior works investigate 3D-aware GANs for 3D consistent multiclass image-to-image (3D-aware 121) translation. Naively using 2D-121 translation methods suffers from unrealistic shape/identity change. To perform 3D-aware multiclass 121 translation, we decouple this learning process into a multiclass 3D-aware GAN step and a 3D-aware 121 translation step. In the first step, we propose two novel techniques: a new conditional architecture and an effective training strategy. In the second step, based on the well-trained multiclass 3D-aware GAN architecture, that preserves view-consistency, we construct a 3D-aware 121 translation system. To further reduce the view-consistency problems, we propose several new techniques, including a U-net-like adaptor network design, a hierarchical representation constrain and a relative regularization loss. In exten-sive experiments on two datasets, quantitative and qualitative results demonstrate that we successfully perform 3D-aware 121 translation with multi-view consistency. Code is available in 3DI2I.
First Page
12652
Last Page
12662
DOI
10.1109/CVPR52729.2023.01217
Publication Date
8-22-2023
Keywords
3D from single images, Training, Computer vision, Three-dimensional displays, Codes, Computational modeling, Computer architecture, Generative adversarial networks
Recommended Citation
S. Li, J. van de Weijer, Y. Wang, F. S. Khan, M. Liu and J. Yang, "3D-Aware Multi-Class Image-to-Image Translation with NeRFs," 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 2023, pp. 12652-12662, doi: 10.1109/CVPR52729.2023.01217.
Comments
Open Access version from CVF
Archived, thanks to CVF
Uploaded 4 June 2024