Text-Guided Synthesis of Crowd Animation

被引:1
|
作者
Ji, Xuebo [1 ]
Pan, Zherong [2 ]
Gao, Xifeng [2 ]
Pan, Jia [1 ]
机构
[1] Univ Hong Kong, Ctr Transformat Garment Prod TransGP, Hong Kong, Peoples R China
[2] LightSpeed Studios, Seattle, WA USA
关键词
Diffusion Model; Multi-Agent Navigation; Collision Avoidance; Crowd Simulation;
D O I
10.1145/3641519.3657516
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Creating vivid crowd animations is core to immersive virtual environments in digital games. This work focuses on tackling the challenges of the crowd behavior generation problem. Existing approaches are labor-intensive, relying on practitioners to manually craft the complex behavior systems. We propose a machine learning approach to synthesize diversified dynamic crowd animation scenarios for a given environment based on a text description input. We first train two conditional diffusion models that generate text-guided agent distribution fields and velocity fields. Assisted by local navigation algorithms, the fields are then used to control multiple groups of agents. We further employ Large-Language Model (LLM) to canonicalize the general script into a structured sentence for more stable training and better scalability. To train our diffusion models, we devise a constructive method to generate random environments and crowd animations. We show that our trained diffusion models can generate crowd animations for both unseen environments and novel scenario descriptions. Our method paves the way towards automatic generating of crowd behaviors for virtual environments. Code and data for this paper are available at: https://github.com/MLZG/Text-Crowd.git.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Text-Guided Molecule Generation with Diffusion Language Model
    Gong, Haisong
    Liu, Qiang
    Wu, Shu
    Wang, Liang
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 1, 2024, : 109 - 117
  • [32] Target-Free Text-Guided Image Manipulation
    Fan, Wan-Cyuan
    Yang, Cheng-Fu
    Yang, Chiao-An
    Wang, Yu-Chiang Frank
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 588 - 596
  • [33] SEGMENTATION-AWARE TEXT-GUIDED IMAGE MANIPULATION
    Haruyama, Tomoki
    Togo, Ren
    Maeda, Keisuke
    Ogawa, Takahiro
    Haseyama, Miki
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 2433 - 2437
  • [34] Prior-free Guided TTS: An Improved and Efficient Diffusion-based Text-Guided Speech Synthesis
    Choi, Won-Gook
    Kim, So-Jeong
    Kim, Taeho
    Chang, Joon-Hyuk
    INTERSPEECH 2023, 2023, : 4289 - 4293
  • [35] Text-Guided Diverse Image Synthesis for Long-Tailed Remote Sensing Object Classification
    Tang, Haojun
    Zhao, Wenda
    Hu, Guang
    Xiao, Yi
    Li, Yunlong
    Wang, Haipeng
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [36] FusionDeformer: text-guided mesh deformation using diffusion models
    Xu, Hao
    Wu, Yiqian
    Tang, Xiangjun
    Zhang, Jing
    Zhang, Yang
    Zhang, Zhebin
    Li, Chen
    Jin, Xiaogang
    VISUAL COMPUTER, 2024, 40 (07): : 4701 - 4712
  • [37] Text-Guided Visual Feature Refinement for Text-Based Person Search
    Gao, Liying
    Niu, Kai
    Ma, Zehong
    Jiao, Bingliang
    Tan, Tonghao
    Wang, Peng
    PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR '21), 2021, : 118 - 126
  • [38] MMFL: Multimodal Fusion Learning for Text-Guided Image Inpainting
    Lin, Qing
    Yan, Bo
    Li, Jichun
    Tan, Weimin
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1094 - 1102
  • [39] Adversarial Learning with Mask Reconstruction for Text-Guided Image Inpainting
    Wu, Xingcai
    Xie, Yucheng
    Zeng, Jiaqi
    Yang, Zhenguo
    Yu, Yi
    Li, Qing
    Liu, Wenyin
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 3464 - 3472
  • [40] Rethinking Super-Resolution as Text-Guided Details Generation
    Ma, Chenxi
    Yan, Bo
    Lin, Qing
    Tan, Weimin
    Chen, Siming
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3461 - 3469