Polyphonic music generation generative adversarial network with Markov decision process

被引:3
|
作者
Huang, Wenkai [1 ]
Xue, Yihao [2 ]
Xu, Zefeng [2 ]
Peng, Guanglong [2 ]
Wu, Yu [3 ]
机构
[1] Guangzhou Univ, Ctr Res Leading Technol Special Equipment, Sch Mech & Elect Engn, Guangzhou 510006, Peoples R China
[2] Guangzhou Univ, Sch Mech & Elect Engn, Guangzhou 510006, Peoples R China
[3] Guangzhou Univ, Lab Ctr, Guangzhou 510006, Peoples R China
关键词
Polyphonic music generation; Markov decision process (MDP); Monte Carlo tree search (MCTS); Deep learning; Wasserstein Generative Adversarial Network (WGAN);
D O I
10.1007/s11042-022-12925-w
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In the process of polyphonic music creation, it is important to combine two or more independent melodies through technical treatment. However, due to the diversity of polyphonic music sequences and the limitations of neural networks, it is difficult to create chords or melodies beyond the training data. As the music sequence increases, the probability of the generator producing the same note will increase, which will destroy the coherence of the music. Therefore, this paper proposes a novel polyphonic music creation model, combining the ideas of the Markov decision process (MDP) and Monte Carlo tree search (MCTS) and improving the Wasserstein Generative Adversarial Network (WGAN) theory. Through the zero-sum game and conditional constraints between generator and discriminator, the model in this study is closer to the unconstrained creation of music, and the growth of music sequence will not affect music coherence. Experimental results show that the algorithm proposed here has a better effect on polyphonic music generation than the latest methods.
引用
收藏
页码:29865 / 29885
页数:21
相关论文
共 50 条
  • [1] Polyphonic music generation generative adversarial network with Markov decision process
    Wenkai Huang
    Yihao Xue
    Zefeng Xu
    Guanglong Peng
    Yu Wu
    [J]. Multimedia Tools and Applications, 2022, 81 : 29865 - 29885
  • [2] Generation of Music With Dynamics Using Deep Convolutional Generative Adversarial Network
    Toh, Raymond Kwan How
    Sourin, Alexei
    [J]. 2021 INTERNATIONAL CONFERENCE ON CYBERWORLDS (CW 2021), 2021, : 137 - 140
  • [3] A transformer generative adversarial network for multi-track music generation
    Jin, Cong
    Wang, Tao
    Li, Xiaobing
    Tie, Chu Jie Jiessie
    Tie, Yun
    Liu, Shan
    Yan, Ming
    Li, Yongzhi
    Wang, Junxian
    Huang, Shenze
    [J]. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2022, 7 (03) : 369 - 380
  • [4] Speech Generation by Generative Adversarial Network
    Chen, Yijia
    [J]. 2021 2ND INTERNATIONAL CONFERENCE ON BIG DATA & ARTIFICIAL INTELLIGENCE & SOFTWARE ENGINEERING (ICBASE 2021), 2021, : 435 - 438
  • [5] Multi-category MIDI music generation based on LSTM Generative adversarial network
    Wang, Yutian
    Yu, Guochen
    Cai, Juanjuan
    Wang, Hui
    [J]. 2019 2ND INTERNATIONAL CONFERENCE ON MECHANICAL, ELECTRONIC AND ENGINEERING TECHNOLOGY (MEET 2019), 2019, : 20 - 25
  • [6] Targeted Speech Adversarial Example Generation With Generative Adversarial Network
    Wang, Donghua
    Dong, Li
    Wang, Rangding
    Yan, Diqun
    Wang, Jie
    [J]. IEEE ACCESS, 2020, 8 : 124503 - 124513
  • [7] Music Creation Technology Based on Generative Adversarial Network
    Liu, Feng
    [J]. JOURNAL OF ELECTRICAL SYSTEMS, 2024, 20 (09) : 626 - 632
  • [8] A novel image semantic communication method via dynamic decision generation network and generative adversarial network
    Liu, Shugang
    Peng, Zhan
    Yu, Qiangguo
    Duan, Linan
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01):
  • [9] MusicGAIL: A Generative Adversarial Imitation Learning Approach for Music Generation
    Liao, Yusong
    Xu, Hongguang
    Xu, Ke
    [J]. ARTIFICIAL INTELLIGENCE, CICAI 2023, PT I, 2024, 14473 : 505 - 516
  • [10] Generative model based polyphonic music transcription
    Cemgil, AT
    Kappen, B
    Barber, D
    [J]. 2003 IEEE WORKSHOP ON APPLICATIONS OF SIGNAL PROCESSING TO AUDIO AND ACOUSTICS PROCEEDINGS, 2003, : 181 - 184