Motion to Dance Music Generation using Latent Diffusion Model

被引:2
|
作者
Tan, Vanessa [1 ]
Nam, JungHyun [1 ]
Nam, Juhan [1 ]
Noh, Junyong [1 ]
机构
[1] KAIST GSCT, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
3D motion to music; music generation; latent diffusion model;
D O I
10.1145/3610543.3626164
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The role of music in games and animation, particularly in dance content, is essential for creating immersive and entertaining experiences. Although recent studies have made strides in generating dance music from videos, their practicality in integrating music into games and animation remains limited. In this context, we present a method capable of generating plausible dance music from 3D motion data and genre labels. Our approach leverages a combination of a UNET-based latent diffusion model and a pre-trained VAE model. To evaluate the performance of the proposed model, we employ evaluation metrics to assess various audio properties, including beat alignment, audio quality, motion-music correlation, and genre score. The quantitative results show that our approach outperforms previous methods. Furthermore, we demonstrate that our model can generate audio that seamlessly fits to in-the-wild motion data. This capability enables us to create plausible dance music that complements dynamic movements of characters and enhances overall audiovisual experience in interactive media. Examples from our proposed model are available at this link: https://dmdproject.github.io/.
引用
收藏
页数:4
相关论文
共 50 条
  • [1] Dance2Music-Diffusion: leveraging latent diffusion models for music generation from dance videos
    Zhang, Chaoyang
    Hua, Yan
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2024, 2024 (01):
  • [2] DiffDance: Cascaded Human Motion Diffusion Model for Dance Generation
    Qi, Qiaosong
    Zhuo, Le
    Zhang, Aixi
    Liao, Yue
    Fang, Fei
    Liu, Si
    Yan, Shuicheng
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 1374 - 1382
  • [3] LaMoD: Latent Motion Diffusion Model for Myocardial Strain Generation
    Xing, Jiarui
    Jayakumar, Nivetha
    Wu, Nian
    Wang, Yu
    Epstein, Frederick H.
    Zhang, Miamniao
    SHAPE IN MEDICAL IMAGING, SHAPEMI 2024, 2025, 15275 : 164 - 177
  • [4] Discrete diffusion model with contrastive learning for music to natural and long dance generation
    Huaxin Wang
    Yujian Jiang
    Xiangzhong Zhou
    Wei Jiang
    npj Heritage Science, 13 (1):
  • [5] Perceptually motivated automatic dance motion generation for music
    Kim, Jae Woo
    Fouad, Hesham
    Sibert, John L.
    Hahn, James K.
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2009, 20 (2-3) : 375 - 384
  • [6] LDMME: Latent Diffusion Model for Music Editing
    Ye, Runchuan
    Kang, Shiyin
    Wu, Zhiyong
    MAN-MACHINE SPEECH COMMUNICATION, NCMMSC 2024, 2025, 2312 : 311 - 325
  • [7] Bidirectional Autoregressive Diffusion Model for Dance Generation
    Zhang, Canyu
    Tang, Youbao
    Zhang, Ning
    Lin, Ruei-Sung
    Han, Mei
    Xiao, Jing
    Wang, Song
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 687 - 696
  • [8] Sign Motion Generation by Motion Diffusion Model
    Hakozaki, Kohei
    Murakami, Tomoya
    Uchida, Tsubasa
    Miyazaki, Taro
    Kaneko, Hiroyuki
    PROCEEDINGS OF THE SIGGRAPH 2024 POSTERS, 2024,
  • [9] A Latent Diffusion Model for Protein Structure Generation
    Fu, Cong
    Yan, Keqiang
    Wang, Limei
    Au, Wing Yee
    McThrow, Michael
    Komikado, Tao
    Maruhashi, Koji
    Uchino, Kanji
    Qian, Xiaoning
    Ji, Shuiwang
    LEARNING ON GRAPHS CONFERENCE, VOL 231, 2023, 231
  • [10] Music to Dance: Motion Generation Based on Multi-Feature Fusion Strategy
    Gao, Yufei
    Yu, Wenxin
    Zhang, Xuewen
    Deng, Xin
    Zhang, Zhiqiang
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 2735 - 2739