TG-Dance: TransGAN-Based Intelligent Dance Generation with Music

被引:0
|
作者
Huang, Dongjin [1 ]
Zhang, Yue [1 ]
Li, Zhenyan [1 ]
Liu, Jinhua [1 ]
机构
[1] Shanghai Univ, Shanghai Film Acad, Shanghai, Peoples R China
来源
关键词
Dance motion generation; Multimodal fusion; Upsampling; Transformer; Multi-head attention;
D O I
10.1007/978-3-031-27077-2_19
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Intelligent choreographic from music is a popular field of study currently. Many works use fragment splicing to generate new motions, which lacks motion diversity. When the input is only music, the frame-by-frame generation methods lead to similar motions generated by the same music. Some works improve this problem by adding motions as one of the inputs, but requires a high number of frames. In this paper, a new transformer-based neural network, TG-dance, is proposed for predicting high-quality 3D dance motions that follow the musical rhythms. We propose a new idea of multi-level expansion of motion sequences and design a new motion encoder, using a multi-level transformer-upsampling layer. The multi-head attention in the transformer allows better access to contextual information. The upsampling can greatly reduce motion frames input, and is memory friendly. We use generative adversarial network to effectively improve the quality of generated motions. We designed experiments on the publicly available large dataset AIST++. The experimental results show that TG-dance network outperforms the latest models in quantitative and qualitative. Our model inputs fewer frames of motion sequences and audio features to predict high-quality 3D dance motion sequences that follow the rhythm of the music.
引用
收藏
页码:243 / 254
页数:12
相关论文
共 50 条
  • [1] THE REMIX GENERATION + DANCE MUSIC
    WATNEY, S
    [J]. ARTFORUM, 1994, 33 (02): : 15 - 16
  • [2] Music2Dance: DanceNet for Music-Driven Dance Generation
    Zhuang, Wenlin
    Wang, Congyi
    Chai, Jinxiang
    Wang, Yangang
    Shao, Ming
    Xia, Siyu
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (02)
  • [3] Robot Dance Generation with Music Based Trajectory Optimization
    Boukheddimi, Melya
    Harnack, Daniel
    Kumar, Shivesh
    Kumar, Rohit
    Vyas, Shubham
    Arriaga, Octavio
    Kirchner, Frank
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 3069 - 3076
  • [4] Music-Driven Dance Generation
    Qi, Yu
    Liu, Yazhou
    Sun, Quansen
    [J]. IEEE ACCESS, 2019, 7 : 166540 - 166550
  • [5] The Collection and Recognition Method of Music and Dance Movement Based on Intelligent Sensor
    Geng, Jun
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [6] EDGE: Editable Dance Generation From Music
    Tseng, Jonathan
    Castellon, Rodrigo
    Liu, C. Karen
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 448 - 458
  • [7] Music-to-Dance Generation with Multiple Conformer
    Zhang, Mingao
    Liu, Changhong
    Chen, Yong
    Lei, Zhenchun
    Wang, Mingwen
    [J]. PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2022, 2022, : 34 - 38
  • [8] Soul Train: The Music, Dance, and Style of a Generation
    Weisbard, Eric
    [J]. AMERICAN QUARTERLY, 2015, 67 (01) : 254 - 265
  • [9] SOUL TRAIN The Music, Dance, and Style of a Generation
    McNatt, Rosemary Bray
    [J]. NEW YORK TIMES BOOK REVIEW, 2014, 119 (22): : 40 - 40
  • [10] Dance2Music-Diffusion: leveraging latent diffusion models for music generation from dance videos
    Zhang, Chaoyang
    Hua, Yan
    [J]. Eurasip Journal on Audio, Speech, and Music Processing, 2024, 2024 (01)