Learning Multi-Task Transferable Rewards via Variational Inverse Reinforcement Learning

被引:0
|
作者
Yoo, Se-Wook [1 ]
Seo, Seung-Woo [1 ]
机构
[1] Seoul Natl Univ, Dept Elect & Comp Engn, Seoul, South Korea
关键词
D O I
10.1109/ICRA.46639.2022.9811697
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Many robotic tasks are composed of a lot of temporally correlated sub-tasks in a highly complex environment. It is important to discover situational intentions and proper actions by deliberating on temporal abstractions to solve problems effectively. To understand the intention separated from changing task dynamics, we extend an empowerment-based regularization technique to situations with multiple tasks based on the framework of a generative adversarial network. Under the multitask environments with unknown dynamics, we focus on learning a reward and policy from the unlabeled expert examples. In this study, we define situational empowerment as the maximum of mutual information representing how an action conditioned on both a certain state and sub-task affects the future. Our proposed method derives the variational lower bound of the situational mutual information to optimize it. We simultaneously learn the transferable multi-task reward function and policy by adding an induced term to the objective function. By doing so, the multi-task reward function helps to learn a robust policy for environmental change. We validate the advantages of our approach on multi-task learning and multitask transfer learning. We demonstrate our proposed method has the robustness of both randomness and changing task dynamics. Finally, we prove that our method has significantly better performance and data efficiency than existing imitation learning methods on various benchmarks.
引用
收藏
页数:7
相关论文
共 50 条
  • [41] HIERARCHICAL MULTI-TASK LEARNING VIA TASK AFFINITY GROUPINGS
    Srivastava, Siddharth
    Bhugra, Swati
    Kaushik, Vinay
    Lall, Brejesh
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 3289 - 3293
  • [42] Leveraging the efficiency of multi-task robot manipulation via task-evoked planner and reinforcement learning
    Qi, Haofu
    Zheng, Haoyang
    Shao, Jun
    Zhang, Jiatao
    Gu, Jason
    Song, Wei
    Zhu, Shiqiang
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 9220 - 9226
  • [43] Multi-task gradient descent for multi-task learning
    Lu Bai
    Yew-Soon Ong
    Tiantian He
    Abhishek Gupta
    Memetic Computing, 2020, 12 : 355 - 369
  • [44] Multi-task gradient descent for multi-task learning
    Bai, Lu
    Ong, Yew-Soon
    He, Tiantian
    Gupta, Abhishek
    MEMETIC COMPUTING, 2020, 12 (04) : 355 - 369
  • [45] Provably Efficient Multi-Task Reinforcement Learning with Model Transfer
    Zhang, Chicheng
    Wang, Zhi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [46] Multi-task Reinforcement Learning in Partially Observable Stochastic Environments
    Li, Hui
    Liao, Xuejun
    Carin, Lawrence
    JOURNAL OF MACHINE LEARNING RESEARCH, 2009, 10 : 1131 - 1186
  • [47] Curriculum-Based Asymmetric Multi-Task Reinforcement Learning
    Huang, Hanchi
    Ye, Deheng
    Shen, Li
    Liu, Wei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (06) : 7258 - 7269
  • [48] Multi-Task Deep Reinforcement Learning for Continuous Action Control
    Yang, Zhaoyang
    Merrick, Kathryn
    Abbass, Hussein
    Jin, Lianwen
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3301 - 3307
  • [49] Multi-task Deep Reinforcement Learning for IoT Service Selection
    Matsuoka, Hiroki
    Moustafa, Ahmed
    ICAART: PROCEEDINGS OF THE 14TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 3, 2022, : 548 - 554
  • [50] Multi-Task Reinforcement Learning with Context-based Representations
    Sodhani, Shagun
    Zhang, Amy
    Pineau, Joelle
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139