Rogue-Gym: A New Challenge for Generalization in Reinforcement Learning

被引:0
|
作者
Kanagawa, Yuji [1 ]
Kaneko, Tomoyuki [2 ]
机构
[1] Univ Tokyo, Grad Sch Arts & Sci, Tokyo, Japan
[2] Univ Tokyo, Interfac Initiat Informat Studies, Tokyo, Japan
关键词
roguelike games; reinforcement learning; generalization; domain adaptation; neural networks; ENVIRONMENT;
D O I
10.1109/cig.2019.8848075
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose Rogue-Gym, a simple and classic style roguelike game built for evaluating generalization in reinforcement learning (RL). Combined with the recent progress of deep neural networks, RL has successfully trained human-level agents without human knowledge in many games such as those for Atari 2600. However, it has been pointed out that agents trained with RL methods often overfit the training environment, and they work poorly in slightly different environments. To investigate this problem, some research environments with procedural content generation have been proposed. Following these studies, we propose the use of roguelikes as a benchmark for evaluating the generalization ability of RL agents. In our Rogue-Gym, agents need to explore dungeons that are structured differently each time they start a new game. Thanks to the very diverse structures of the dungeons, we believe that the generalization benchmark of Rogue-Gym is sufficiently fair. In our experiments, we evaluate a standard reinforcement learning method, PPO, with and without enhancements for generalization. The results show that some enhancements believed to be effective fail to mitigate the overfitting in Rogue-Gym, although others slightly improve the generalization ability.
引用
下载
收藏
页数:8
相关论文
共 50 条
  • [21] Gym Hero: A Research Environment for Reinforcement Learning Agents in Rhythm Games
    Ferrer Filho, Romulo Freire
    Barbosa Nogueira, Yuri Lenon
    Vidal, Creto Augusto
    Cavalcante-Neto, Joaquim Bento
    de Sousa Serafim, Paulo Bruno
    2021 20TH BRAZILIAN SYMPOSIUM ON COMPUTER GAMES AND DIGITAL ENTERTAINMENT (SBGAMES 2021), 2021, : 87 - 96
  • [22] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Krishnan, Srivatsan
    Boroujerdian, Behzad
    Fu, William
    Faust, Aleksandra
    Reddi, Vijay Janapa
    MACHINE LEARNING, 2021, 110 (09) : 2501 - 2540
  • [23] Air Learning: a deep reinforcement learning gym for autonomous aerial robot visual navigation
    Srivatsan Krishnan
    Behzad Boroujerdian
    William Fu
    Aleksandra Faust
    Vijay Janapa Reddi
    Machine Learning, 2021, 110 : 2501 - 2540
  • [24] Abstraction and Generalization in Reinforcement Learning: A Summary and Framework
    Ponsen, Marc
    Taylor, Matthew E.
    Tuyls, Karl
    ADAPTIVE AND LEARNING AGENTS, 2010, 5924 : 1 - +
  • [25] Towards Min Max Generalization in Reinforcement Learning
    Fonteneau, Raphael
    Murphy, Susan A.
    Wehenkel, Louis
    Ernst, Damien
    AGENTS AND ARTIFICIAL INTELLIGENCE, 2011, 129 : 61 - +
  • [26] Decoupling Value and Policy for Generalization in Reinforcement Learning
    Raileanu, Roberta
    Fergus, Rob
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [27] Generalization in Reinforcement Learning by Soft Data Augmentation
    Hansen, Nicklas
    Wang, Xiaolong
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 13611 - 13617
  • [28] Automatic Data Augmentation for Generalization in Reinforcement Learning
    Raileanu, Roberta
    Goldstein, Max
    Yarats, Denis
    Kostrikov, Ilya
    Fergus, Rob
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [29] Instance-based Generalization in Reinforcement Learning
    Bertran, Martin
    Martinez, Natalia
    Phielipp, Mariano
    Sapiro, Guillermo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [30] Improving Generalization in Reinforcement Learning with Mixture Regularization
    Wang, Kaixin
    Kang, Bingyi
    Shao, Jie
    Feng, Jiashi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33