Motion to Dance Music Generation using Latent Diffusion Model

被引:2
|
作者
Tan, Vanessa [1 ]
Nam, JungHyun [1 ]
Nam, Juhan [1 ]
Noh, Junyong [1 ]
机构
[1] KAIST GSCT, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
3D motion to music; music generation; latent diffusion model;
D O I
10.1145/3610543.3626164
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The role of music in games and animation, particularly in dance content, is essential for creating immersive and entertaining experiences. Although recent studies have made strides in generating dance music from videos, their practicality in integrating music into games and animation remains limited. In this context, we present a method capable of generating plausible dance music from 3D motion data and genre labels. Our approach leverages a combination of a UNET-based latent diffusion model and a pre-trained VAE model. To evaluate the performance of the proposed model, we employ evaluation metrics to assess various audio properties, including beat alignment, audio quality, motion-music correlation, and genre score. The quantitative results show that our approach outperforms previous methods. Furthermore, we demonstrate that our model can generate audio that seamlessly fits to in-the-wild motion data. This capability enables us to create plausible dance music that complements dynamic movements of characters and enhances overall audiovisual experience in interactive media. Examples from our proposed model are available at this link: https://dmdproject.github.io/.
引用
收藏
页数:4
相关论文
共 50 条
  • [41] Multisite precipitation generation using a latent autoregressive model
    Rasmussen, P. F.
    WATER RESOURCES RESEARCH, 2013, 49 (04) : 1845 - 1857
  • [42] Dance-the-Music: an educational platform for the modeling, recognition and audiovisual monitoring of dance steps using spatiotemporal motion templates
    Pieter-Jan Maes
    Denis Amelynck
    Marc Leman
    EURASIP Journal on Advances in Signal Processing, 2012
  • [43] A deep learning model of dance generation for young children based on music rhythm and beat
    Kong, Shanshan
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (13):
  • [44] Dance-the-Music: an educational platform for the modeling, recognition and audiovisual monitoring of dance steps using spatiotemporal motion templates
    Maes, Pieter-Jan
    Amelynck, Denis
    Leman, Marc
    EURASIP JOURNAL ON ADVANCES IN SIGNAL PROCESSING, 2012,
  • [45] DanceConv: Dance Motion Generation With Convolutional Networks
    Kritsis, Kosmas
    Gkiokas, Aggelos
    Pikrakis, Aggelos
    Katsouros, Vassilis
    IEEE ACCESS, 2022, 10 : 44982 - 45000
  • [46] DanceConv: Dance Motion Generation With Convolutional Networks
    Kritsis, Kosmas
    Gkiokas, Aggelos
    Pikrakis, Aggelos
    Katsouros, Vassilis
    IEEE Access, 2022, 10 : 44982 - 45000
  • [47] A human motion analysis using the rhythm - A estimate method of dance motion with autoregressive model
    Kojima, K
    Otobe, T
    Hironaga, M
    Nagae, S
    IEEE RO-MAN 2000: 9TH IEEE INTERNATIONAL WORKSHOP ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, PROCEEDINGS, 2000, : 190 - 193
  • [48] Local Action-Guided Motion Diffusion Model for Text-to-Motion Generation
    Jin, Peng
    Li, Hao
    Cheng, Zesen
    Li, Kehan
    Yu, Runyi
    Liu, Chang
    Ji, Xiangyang
    Yuan, Li
    Chen, Jie
    COMPUTER VISION - ECCV 2024, PT XXV, 2025, 15083 : 392 - 409
  • [49] EMDM: Efficient Motion Diffusion Model for Fast and High-Quality Motion Generation
    Zhou, Wenyang
    Dou, Zhiyang
    Cao, Zeyu
    Liao, Zhouyingcheng
    Wang, Jingbo
    Wang, Wenjia
    Liu, Yuan
    Komura, Taku
    Wang, Wenping
    Liu, Lingjie
    COMPUTER VISION - ECCV 2024, PT II, 2025, 15060 : 18 - 38
  • [50] Diff-BGM: A Diffusion Model for Video Background Music Generation
    Li, Sizhe
    Qin, Yiming
    Zheng, Minghang
    Jin, Xin
    Liu, Yang
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 27338 - 27347