Advances in text-guided 3D editing: a survey

被引:0
|
作者
Lu, Lihua [1 ]
Li, Ruyang [1 ]
Zhang, Xiaohui [1 ]
Wei, Hui [1 ]
Du, Guoguang [1 ]
Wang, Binqiang [1 ]
机构
[1] Shandong Mass Informat Technol Res Inst, Jinan, Peoples R China
关键词
Text prompts; Text-guided 3D editing; Editing capacity; NEURAL RADIANCE FIELDS;
D O I
10.1007/s10462-024-10937-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In 3D Artificial Intelligence Generated Content (AIGC), compared with generating 3D assets from scratch, editing extant 3D assets satisfies user prompts, allowing the creation of diverse and high-quality 3D assets in a time and labor-saving manner. More recently, text-guided 3D editing that modifies 3D assets guided by text prompts is user-friendly and practical, which evokes a surge in research within this field. In this survey, we comprehensively investigate recent literature on text-guided 3D editing in an attempt to answer two questions: What are the methodologies of existing text-guided 3D editing? How has current progress in text-guided 3D editing gone so far? Specifically, we focus on text-guided 3D editing methods published in the past 4 years, delving deeply into their frameworks and principles. We then present a fundamental taxonomy in terms of the editing strategy, optimization scheme, and 3D representation. Based on the taxonomy, we review recent advances in this field, considering factors such as editing scale, type, granularity, and perspective. In addition, we highlight four applications of text-guided 3D editing, including texturing, style transfer, local editing of scenes, and insertion editing, to exploit further the 3D editing capacities with in-depth comparisons and discussions. Depending on the insights achieved by this survey, we discuss open challenges and future research directions. We hope this survey will help readers gain a deeper understanding of this exciting field and foster further advancements in text-guided 3D editing.
引用
收藏
页数:61
相关论文
共 50 条
  • [21] DiffusionGAN3D: Boosting Text-guided 3D Generation and Domain Adaptation by Combining 3D GANs and Diffusion Priors
    Lei, Biwen
    Yu, Kai
    Feng, Mengyang
    Cui, Miaomiao
    Xie, Xuansong
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 10487 - 10497
  • [22] Contrastive Denoising Score for Text-guided Latent Diffusion Image Editing
    Nam, Hyelin
    Kwon, Gihyun
    Park, Geon Yeong
    Ye, Jong Chul
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 9192 - 9201
  • [23] SKED: Sketch-guided Text-based 3D Editing
    Mikaeili, Aryan
    Perel, Or
    Safaee, Mehdi
    Cohen-Or, Daniel
    Mahdavi-Amiri, Ali
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 14561 - 14573
  • [24] CLIP-Head: Text-Guided Generation of Textured Neural Parametric 3D Head Models
    Manu, Pranav
    Srivastava, Astitva
    Sharma, Avinash
    PROCEEDINGS SIGGRAPH ASIA 2023 TECHNICAL COMMUNICATIONS, SA TECHNICAL COMMUNICATIONS 2023, 2023,
  • [25] MFECLIP: CLIP With Mapping-Fusion Embedding for Text-Guided Image Editing
    Wu, Fei
    Ma, Yongheng
    Jin, Hao
    Jing, Xiao-Yuan
    Jiang, Guo-Ping
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 116 - 120
  • [26] GLIDE: Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models
    Nichol, Alex
    Dhariwal, Prafulla
    Ramesh, Aditya
    Shyam, Pranav
    Mishkin, Pamela
    McGrew, Bob
    Sutskever, Ilya
    Chen, Mark
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [27] Controlling Attention Map Better for Text-Guided Image Editing Diffusion Models
    Xu, Siqi
    Sun, Lijun
    Liu, Guanming
    Wei, Zhihua
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT XII, ICIC 2024, 2024, 14873 : 54 - 65
  • [28] DE-net: Dynamic Text-Guided Image Editing Adversarial Networks
    Tao, Ming
    Bao, Bing-Kun
    Tang, Hao
    Wu, Fei
    Wei, Longhui
    Tian, Qi
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 8, 2023, : 9971 - 9979
  • [29] AnimatableDreamer: Text-Guided Non-rigid 3D Model Generation and Reconstruction with Canonical Score Distillation
    Wang, Xinzhou
    Wang, Yikai
    Yee, Junliang
    Sung, Fuchun
    Wang, Zhengyi
    Wang, Ling
    Liu, Pengkun
    Sung, Kai
    Wan, Xintong
    Xie, Wende
    Liu, Fangfu
    He, Bin
    COMPUTER VISION - ECCV 2024, PT XXV, 2025, 15083 : 321 - 339
  • [30] Towards High-Fidelity Text-Guided 3D Face Generation and Manipulation Using only Images
    Yu, Cuican
    Lu, Guansong
    Zeng, Yihan
    Sun, Jian
    Liang, Xiaodan
    Li, Huibin
    Xu, Zongben
    Xu, Songcen
    Zhang, Wei
    Xu, Hang
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 15280 - 15291