GIRAFFE HD: A High-Resolution 3D-aware Generative Model (original) (raw)
Yang Xue1 | Yuheng Li2 | Krishna Kumar Singh3 | Yong Jae Lee2 |
---|
1UC Davis | 2UW–Madison | 3Adobe Research |
---|
Abstract
3D-aware generative models have shown that the introduction of 3D information can lead to more controllable image generation. In particular, the current state-of-the-art model GIRAFFE [M. Niemeyer and A. Geiger, CVPR 2021] can control each object's rotation, translation, scale, and scene camera pose without corresponding supervision. However, GIRAFFE only operates well when the image resolution is low. We propose GIRAFFE HD, a high-resolution 3D-aware generative model that inherits all of GIRAFFE's controllable features while generating high-quality, high-resolution images (5122 resolution and above). The key idea is to leverage a style-based neural renderer, and to independently generate the foreground and background to force their disentanglement while imposing consistency constraints to stitch them together to composite a coherent final image. We demonstrate state-of-the-art 3D controllable high-resolution image generation on multiple natural image datasets.
Architecture
Controllable Scene Generation
GIRAFFE HD preserves all of GIRAFFE's 3D controllable features while generating images at significantly higher quality and resolution. Importantly, compared to GIRAFFE, our background remains more consistent when the foreground changes. Below we show samples of AFHQ Cat2562, CompCar 5122 and FFHQ 10242.
Object Appearance
Background
Object Shape
Camera Elevation
Translation
Scaling
Multiple objects
Rotation
Citation
@inproceedings{xue2022giraffehd,
author = {Yang Xue and Yuheng Li and Krishna Kumar Singh and Yong Jae Lee},
title = {GIRAFFE HD: A High-Resolution 3D-aware Generative Model},
booktitle = {CVPR},
year = {2022},
}