BlendNeRF - Official PyTorch Implementation

3D-aware Blending with Generative NeRFs
Hyunsu Kim1, Gayoung Lee1, Yunjey Choi1, Jin-Hwa Kim1,2, Jun-Yan Zhu3
1NAVER AI Lab, 2SNU AIIS, 3CMU

Project page | Arxiv

Abstract: Image blending aims to combine multiple images seamlessly. It remains challenging for existing 2D-based methods, especially when input images are misaligned due to differences in 3D camera poses and object shapes. To tackle these issues, we propose a 3D-aware blending method using generative Neural Radiance Fields (NeRF), including two key components: 3D-aware alignment and 3D-aware blending. For 3D-aware alignment, we first estimate the camera pose of the reference image with respect to generative NeRFs and then perform 3D local alignment for each part. To further leverage 3D information of the generative NeRF, we propose 3D-aware blending that directly blends images on the NeRF's latent representation space, rather than raw pixel space. Collectively, our method outperforms existing 2D baselines, as validated by extensive quantitative and qualitative evaluations with FFHQ and AFHQ-Cat.

Code will come soon!

Multi-view blending results

CelebA-HQ (FFHQ-pretrained EG3D)

AFHQv2-Cat (EG3D)

ShapeNet-Car (EG3D)

FFHQ (StyleSDF)

Comparison with baselines

CelebA-HQ

AFHQv2-Cat

Citation

If you find this work useful for your research, please cite our paper:

@article{kim20233daware,
    title={3D-aware Blending with Generative NeRFs}, 
    author={Hyunsu Kim and Gayoung Lee and Yunjey Choi and Jin-Hwa Kim and Jun-Yan Zhu},
    journal={arXiv preprint arXiv:2302.06608},
    year={2023}
}