ROBOPIANIST: Dexterous Piano Playing with Deep Reinforcement Learning

被引:0
|
作者
Zakka, Kevin [1 ,2 ]
Wu, Philipp [1 ]
Smith, Laura [1 ]
Gileadi, Nimrod [2 ]
Howell, Taylor [3 ]
Peng, Xue Bin [4 ]
Singh, Sumeet [2 ]
Tassa, Yuval [2 ]
Florence, Pete [2 ]
Zeng, Andy [2 ]
Abbeel, Pieter [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
[2] Google DeepMind, London, England
[3] Stanford Univ, Stanford, CA 94305 USA
[4] Simon Fraser Univ, Burnaby, BC, Canada
来源
关键词
high-dimensional control; bi-manual dexterity; MANIPULATION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Replicating human-like dexterity in robot hands represents one of the largest open problems in robotics. Reinforcement learning is a promising approach that has achieved impressive progress in the last few years; however, the class of problems it has typically addressed corresponds to a rather narrow definition of dexterity as compared to human capabilities. To address this gap, we investigate piano-playing, a skill that challenges even the human limits of dexterity, as a means to test high-dimensional control, and which requires high spatial and temporal precision, and complex finger coordination and planning. We introduce ROBOPIANIST, a system that enables simulated anthropomorphic hands to learn an extensive repertoire of 150 piano pieces where traditional model-based optimization struggles. We additionally introduce an open-sourced environment, benchmark of tasks, interpretable evaluation metrics, and open challenges for future study. Our website featuring videos, code, and datasets is available at https://kzakka.com/robopianist/.
引用
收藏
页数:20
相关论文
共 50 条
  • [1] An investigation on the application of deep reinforcement learning in piano playing technique training
    Ji, Chen
    Wang, Dan
    Wang, Huan
    [J]. Applied Mathematics and Nonlinear Sciences, 2024, 9 (01)
  • [2] Learning Complex Dexterous Manipulation with Deep Reinforcement Learning and Demonstrations
    Rajeswaran, Aravind
    Kumar, Vikash
    Gupta, Abhishek
    Vezzani, Giulia
    Schulman, John
    Todorov, Emanuel
    Levine, Sergey
    [J]. ROBOTICS: SCIENCE AND SYSTEMS XIV, 2018,
  • [3] Playing FPS Games with Deep Reinforcement Learning
    Lample, Guillaume
    Chaplot, Devendra Singh
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2140 - 2146
  • [4] Deep Reinforcement Learning for General Game Playing
    Goldwaser, Adrian
    Thielscher, Michael
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 1701 - 1708
  • [5] PIANO: Influence Maximization Meets Deep Reinforcement Learning
    Li, Hui
    Xu, Mengting
    Bhowmick, Sourav S.
    Rayhan, Joty Shafiq
    Sun, Changsheng
    Cui, Jiangtao
    [J]. IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2023, 10 (03) : 1288 - 1300
  • [6] Deep Reinforcement Learning for Conversational Robots Playing Games
    Cuayahuitl, Heriberto
    [J]. 2017 IEEE-RAS 17TH INTERNATIONAL CONFERENCE ON HUMANOID ROBOTICS (HUMANOIDS), 2017, : 771 - 776
  • [7] Dexterous Manipulation with Deep Reinforcement Learning: Efficient, General, and Low-Cost
    Zhu, Henry
    Gupta, Abhishek
    Rajeswaran, Aravind
    Levine, Sergey
    Kumar, Vikash
    [J]. 2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 3651 - 3657
  • [8] Real-world dexterous object manipulation based deep reinforcement learning
    Yao, Qingfeng
    Wang, Jilong
    Yang, Shuyu
    [J]. arXiv, 2021,
  • [9] Deep reinforcement learning in playing Tetris with robotic arm experiment
    Yan, Yu
    Liu, Peng
    Zhao, Jin
    Zhang, Chengxi
    Wang, Guangwei
    [J]. TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2022,
  • [10] Comparison of Deep Reinforcement Learning Approaches for Intelligent Game Playing
    Jeerige, Anoop
    Bein, Doina
    Verma, Abhishek
    [J]. 2019 IEEE 9TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2019, : 366 - 371