DiffusionGAN3D: Boosting Text-guided 3D Generation and Domain Adaptation by Combining 3D GANs and Diffusion Priors

被引:0
|
作者
Lei, Biwen [1 ]
Yu, Kai [1 ]
Feng, Mengyang [1 ]
Cui, Miaomiao [1 ]
Xie, Xuansong [1 ]
机构
[1] Alibaba Grp, Hangzhou, Peoples R China
关键词
D O I
10.1109/CVPR52733.2024.00998
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text-guided domain adaptation and generation of 3D-aware portraits find many applications in various fields. However, due to the lack of training data and the challenges in handling the high variety of geometry and appearance, the existing methods for these tasks suffer from issues like inflexibility, instability, and low fidelity. In this paper, we propose a novel framework DiffusionGAN3D, which boosts text-guided 3D domain adaptation and generation by combining 3D GANs and diffusion priors. Specifically, we integrate the pre-trained 3D generative models (e.g., EG3D) and text-to-image diffusion models. The former provides a strong foundation for stable and high-quality avatar generation from text. And the diffusion models in turn offer powerful priors and guide the 3D generator finetuning with informative direction to achieve flexible and efficient text-guided domain adaptation. To enhance the diversity in do-main adaptation and the generation capability in text-to-avatar, we introduce the relative distance loss and case-specific learnable triplane respectively. Besides, we design a progressive texture refinement module to improve the texture quality for both tasks above. Extensive experiments demonstrate that the proposed framework achieves excellent results in both domain adaptation and text-to-avatar tasks, outperforming existing methods in terms of generation quality and efficiency. The project homepage is at https://younglbw.github.io/DiffusionGAN3D-homepage/.
引用
收藏
页码:10487 / 10497
页数:11
相关论文
共 50 条
  • [41] AutoSDF: Shape Priors for 3D Completion, Reconstruction and Generation
    Mittal, Paritosh
    Cheng, Yen-Chi
    Singh, Maneesh
    Tulsiani, Shubham
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 306 - 315
  • [42] Improving 3D similarity search by enhancing and combining 3D descriptors
    Bustos, Benjamin
    Schreck, Tobias
    Walter, Michael
    Manuel Barrios, Juan
    Schaefer, Matthias
    Keim, Daniel
    MULTIMEDIA TOOLS AND APPLICATIONS, 2012, 58 (01) : 81 - 108
  • [43] 3D Model Data Generation and Conversion for 3D Printers
    Sungsuk Kim
    Journal of Electronic Materials, 2015, 44 : 815 - 822
  • [44] Automatic 3D object placement for 3D scene generation
    Akazawa, Y
    Okada, Y
    Niijima, K
    MODELLING AND SIMULATION 2003, 2003, : 316 - 318
  • [45] Improving 3D similarity search by enhancing and combining 3D descriptors
    Benjamin Bustos
    Tobias Schreck
    Michael Walter
    Juan Manuel Barrios
    Matthias Schaefer
    Daniel Keim
    Multimedia Tools and Applications, 2012, 58 : 81 - 108
  • [46] 3D Model Data Generation and Conversion for 3D Printers
    Kim, Sungsuk
    JOURNAL OF ELECTRONIC MATERIALS, 2015, 44 (03) : 815 - 822
  • [47] EpiGRAF: Rethinking training of 3D GANs
    Skorokhodov, Ivan
    Tulyakov, Sergey
    Wang, Yiqun
    Wonka, Peter
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [48] Structured Domain Adaptation for 3D Keypoint Estimation
    Vasconcelos, Levi O.
    Mancini, Massimiliano
    Boscaini, Davide
    Caputo, Barbara
    Ricci, Elisa
    2019 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2019), 2019, : 57 - 66
  • [49] Efficient 3D mesh adaptation in diffusion simulation
    Chen, T
    Yergeau, DW
    Dutton, RW
    SISPAD '96 - 1996 INTERNATIONAL CONFERENCE ON SIMULATION OF SEMICONDUCTOR PROCESSES AND DEVICES, 1996, : 171 - 172
  • [50] 3D VR Sketch Guided 3D Shape Prototyping and Exploration
    Luo, Ling
    Chowdhury, Pinaki Nath
    Xiang, Tao
    Song, Yi-Zhe
    Gryaditskaya, Yulia
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 9233 - 9242