Text2Mesh: Text-Driven Neural Stylization for Meshes

被引:120
|
作者
Michel, Oscar [1 ]
Bar-On, Roi [1 ,2 ]
Liu, Richard [1 ]
Benaim, Sagie [2 ]
Hanocka, Rana [1 ]
机构
[1] Univ Chicago, Chicago, IL 60637 USA
[2] Tel Aviv Univ, Tel Aviv, Israel
关键词
D O I
10.1109/CVPR52688.2022.01313
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we develop intuitive controls for editing the style of 3D objects. Our framework, Text2Mesh, stylizes a 3D mesh by predicting color and local geometric details which conform to a target text prompt. We consider a disentangled representation of a 3D object using a fixed mesh input (content) coupled with a learned neural network, which we term a neural style field network (NSF). In order to modify style, we obtain a similarity score between a text prompt (describing style) and a stylized mesh by harnessing the representational power of CLIP. Text2Mesh requires neither a pre-trained generative model nor a specialized 3D mesh dataset. It can handle low-quality meshes (non-manifold, boundaries, etc.) with arbitrary genus, and does not require UV parameterization. We demonstrate the ability of our technique to synthesize a myriad of styles over a wide variety of 3D meshes. Our code and results are available in our project webpage: https://threedle.github.io/text2mesh/.
引用
收藏
页码:13482 / 13492
页数:11
相关论文
共 50 条
  • [41] Shape-aware Text-driven Layered Video Editing
    Lee, Yao-Chih
    Jang, Ji-Ze Genevieve
    Chen, Yi-Ting
    Qiu, Elizabeth
    Huang, Jia-Bin
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 14317 - 14326
  • [42] Lightweight Text-Driven Image Editing With Disentangled Content and Attributes
    Li, Bo
    Lin, Xiao
    Liu, Bin
    He, Zhi-Fen
    Lai, Yu-Kun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 1829 - 1841
  • [43] Semiotic modeling of the text-driven conceptual paradigm in language education
    Li, Yufeng
    CHINESE SEMIOTIC STUDIES, 2021, 17 (04) : 661 - 683
  • [44] Text-driven Emotional Style Control and Cross-speaker Style Transfer in Neural TTS
    Shin, Yookyung
    Lee, Younggun
    Jo, Suhee
    Hwang, Yeongtae
    Kim, Taesu
    INTERSPEECH 2022, 2022, : 2313 - 2317
  • [45] TEXT2VIDEO: TEXT-DRIVEN TALKING-HEAD VIDEO SYNTHESIS WITH PERSONALIZED PHONEME - POSE DICTIONARY
    Zhang, Sibo
    Yuan, Jiahong
    Liao, Miao
    Zhang, Liangjun
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2659 - 2663
  • [46] MotionDiffuse: Text-Driven Human Motion Generation With Diffusion Model
    Zhang, Mingyuan
    Cai, Zhongang
    Pan, Liang
    Hong, Fangzhou
    Guo, Xinying
    Yang, Lei
    Liu, Ziwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (06) : 4115 - 4128
  • [47] TextManiA: Enriching Visual Feature by Text-driven Manifold Augmentation
    Ye-Bin, Moon
    Kim, Jisoo
    Kim, Hongyeob
    Son, Kilho
    Oh, Tae-Hyun
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 2526 - 2537
  • [48] Text-Driven Generative Domain Adaptation with Spectral Consistency Regularization
    Liu, Zhenhuan
    Li, Liang
    Xiao, Jiayu
    Zha, Zheng-Jun
    Huang, Qingming
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 6996 - 7006
  • [49] Text-driven human image generation with texture and pose control
    Jin, Zhedong
    Xia, Guiyu
    Yang, Paike
    Wang, Mengxiang
    Sun, Yubao
    Liu, Qingshan
    NEUROCOMPUTING, 2025, 634
  • [50] Multi-Region Text-Driven Manipulation of Diffusion Imagery
    Li, Yiming
    Zhou, Peng
    Sun, Jun
    Xu, Yi
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 3261 - 3269