ControlNeRF: Text-Driven 3D Scene Stylization via Diffusion Model

被引:0
|
作者
Chen, Jiahui [1 ]
Yang, Chuanfeng [1 ]
Li, Kaiheng [1 ]
Wu, Qingqiang [1 ]
Hong, Qingqi [1 ]
机构
[1] Xiamen Univ, Dept Digital Media Technol, Xiamen, Peoples R China
关键词
Stylization; Neural Radiance Fields; Diffusion Model; View Synthesis;
D O I
10.1007/978-3-031-72335-3_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D scene stylization aims to generate artistically rendered images from various viewpoints within a 3D space while ensuring style consistency regardless of the viewing angle. Traditional 2D methods usually used in this field struggle with maintaining this consistency when applied to 3D environments. To address this issue, we propose a novel approach named ControlNeRF, which employs a customized conditional diffusion model, ControlNet, and introduces latent variables, obtaining a stylized appearance throughout the scene solely driven by text. Specifically, this text-driven approach effectively overcomes the inconveniences associated with using images as style cues, and it not only achieves a high degree of stylistic consistency across various viewpoints but also produces high-quality images. We have conducted rigorous testing on ControlNeRF with diverse styles, which has confirmed these outcomes. Our approach not only advances the field of 3D scene stylization but also opens new possibilities for artistic expression and digital imaging.
引用
收藏
页码:395 / 406
页数:12
相关论文
共 50 条
  • [31] Blended Diffusion for Text-driven Editing of Natural Images
    Avrahami, Omri
    Lischinski, Dani
    Fried, Ohad
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18187 - 18197
  • [32] LGTM: Local-to-Global Text-Driven Human Motion Diffusion Model
    Sun, Haowen
    Zheng, Ruikun
    Huang, Haibin
    Ma, Chongyang
    Huang, Hui
    Hu, Ruizhen
    PROCEEDINGS OF SIGGRAPH 2024 CONFERENCE PAPERS, 2024,
  • [33] StylizedNeRF: Consistent 3D Scene Stylization as Stylized NeRF via 2D-3D Mutual Learning
    Huang, Yi-Hua
    He, Yue
    Yuan, Yu-Jie
    Lai, Yu-Kun
    Gao, Lin
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18321 - 18331
  • [34] FaceCLIPNeRF: Text-driven 3D Face Manipulation using Deformable Neural Radiance Fields
    Hwang, Sungwon
    Hyung, Junha
    Kim, Daejin
    Kim, Min-Jung
    Choo, Jaegul
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 3446 - 3456
  • [35] 3D-Aware Text-Driven Talking Avatar Generation
    Wu, Xiuzhe
    Sun, Yang-Tian
    Chen, Handi
    Zhou, Hang
    Wang, Jingdong
    Liu, Zhengzhe
    Qi, Xiaojuan
    COMPUTER VISION - ECCV 2024, PT LXXXVIII, 2025, 15146 : 416 - 433
  • [36] PNeSM: Arbitrary 3D Scene Stylization via Prompt-Based Neural Style Mapping
    Chen, Jiafu
    Xing, Wei
    Sun, Jiakai
    Chu, Tianyi
    Huang, Yiling
    Ji, Boyan
    Zhao, Lei
    Lin, Huaizhong
    Chen, Haibo
    Wang, Zhizhong
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 2, 2024, : 1091 - 1099
  • [37] WaSt-3D: Wasserstein-2 Distance for Scene-to-Scene Stylization on 3D Gaussians
    Kotovenko, Dmytro
    Grebenkova, Olga
    Sarafianos, Nikolaos
    Paliwal, Avinash
    Ma, Pingchuan
    Poursaeed, Omid
    Mohan, Sreyas
    Fang, Yuchen
    Li, Yilei
    Ranjan, Rakesh
    Ommer, Bjoern
    COMPUTER VISION - ECCV 2024, PT XXI, 2025, 15079 : 298 - 314
  • [38] Multi-Region Text-Driven Manipulation of Diffusion Imagery
    Li, Yiming
    Zhou, Peng
    Sun, Jun
    Xu, Yi
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 3261 - 3269
  • [39] UniTune: Text-Driven Image Editing by Fine Tuning a Diffusion Model on a Single Image
    Valevski, Dani
    Kalman, Matan
    Molad, Eyal
    Segalis, Eyal
    Matias, Yossi
    Leviathan, Yaniv
    ACM TRANSACTIONS ON GRAPHICS, 2023, 42 (04):
  • [40] TexFit: Text-Driven Fashion Image Editing with Diffusion Models
    Wang, Tongxin
    Ye, Mang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 9, 2024, : 10198 - 10206