Anticipatory model of musical style imitation using collaborative and competitive reinforcement learning

被引:0
|
作者
Cont, Arshia [1 ,2 ]
Dubnov, Shlomo [2 ]
Assayag, Gerard [1 ]
机构
[1] Ircam Ctr Pompidou, UMR CNRS 9912, F-9912 Paris, France
[2] Univ Calif San Diego, Ctr Res Comp, San Diego, CA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The role of expectation in listening and composing music has drawn much attention in music cognition since about half a century ago. In this paper, we provide a first attempt to model some aspects of musical expectation specifically pertained to short-time and working memories, in an anticipatory framework. In our proposition anticipation is the mental realization of possible predicted actions and their effect on the perception of the world at an instant in time. We demonstrate the model in applications to automatic improvisation and style imitation. The proposed model, based on cognitive foundations of musical expectation, is an active model using reinforcement learning techniques with multiple agents that learn competitively and in collaboration. We show that compared to similar models, this anticipatory framework needs little training data and demonstrates complex musical behavior such as long-term planning and formal shapes as a result of the anticipatory architecture. We provide sample results and discuss further research.
引用
收藏
页码:285 / +
页数:3
相关论文
共 50 条
  • [1] Generalized Path Planning for Collaborative UAVs using Reinforcement and Imitation Learning
    Farley, Jack
    Chapnevis, Amirahmad
    Bulut, Eyuphan
    PROCEEDINGS OF THE 2023 INTERNATIONAL SYMPOSIUM ON THEORY, ALGORITHMIC FOUNDATIONS, AND PROTOCOL DESIGN FOR MOBILE NETWORKS AND MOBILE COMPUTING, MOBIHOC 2023, 2023, : 457 - 462
  • [2] Cooperative and Competitive Reinforcement and Imitation Learning for a Mixture of Heterogeneous Learning Modules
    Uchibe, Eiji
    FRONTIERS IN NEUROROBOTICS, 2018, 12
  • [3] Learning to Drive Using Sparse Imitation Reinforcement Learning
    Han, Yuci
    Yilmaz, Alper
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 3736 - 3742
  • [4] Imitation Game: A Model-based and Imitation Learning Deep Reinforcement Learning Hybrid
    Veith, Eric Msp
    Logemann, Torben
    Berezin, Aleksandr
    Wellssow, Arlena
    Balduin, Stephan
    2024 12TH WORKSHOP ON MODELING AND SIMULATION OF CYBER-PHYSICAL ENERGY SYSTEMS, MSCPES, 2024,
  • [5] Using reinforcement learning to adapt an imitation task
    Guenter, Florent
    Billard, Aude G.
    2007 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-9, 2007, : 1028 - 1033
  • [6] Collaborative promotion: Achieving safety and task performance by integrating imitation reinforcement learning
    Zhang, Cai
    Zhang, Xiaoxiong
    Zhang, Hui
    Zhu, Fei
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255
  • [7] Generating stable molecules using imitation and reinforcement learning
    Meldgaard, Soren Ager
    Koehler, Jonas
    Mortensen, Henrik Lund
    Christiansen, Mads-Peter, V
    Noe, Frank
    Hammer, Bjork
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (01):
  • [8] Symbolization and imitation learning of motion sequence using competitive modules
    Samejima, K
    Katagiri, K
    Doya, K
    Kawato, M
    ELECTRONICS AND COMMUNICATIONS IN JAPAN PART III-FUNDAMENTAL ELECTRONIC SCIENCE, 2006, 89 (09): : 42 - 53
  • [9] A biologically inspired method for conceptual imitation using reinforcement learning
    Mobahi, Hossein
    Ahmadabadi, Majid Nili
    Araabi, Babak Nadjar
    APPLIED ARTIFICIAL INTELLIGENCE, 2007, 21 (03) : 155 - 183
  • [10] ReAgent: Point Cloud Registration using Imitation and Reinforcement Learning
    Bauer, Dominik
    Patten, Timothy
    Vincze, Markus
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 14581 - 14589