Learning Multi-Task Transferable Rewards via Variational Inverse Reinforcement Learning

被引:0
|
作者
Yoo, Se-Wook [1 ]
Seo, Seung-Woo [1 ]
机构
[1] Seoul Natl Univ, Dept Elect & Comp Engn, Seoul, South Korea
关键词
D O I
10.1109/ICRA.46639.2022.9811697
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Many robotic tasks are composed of a lot of temporally correlated sub-tasks in a highly complex environment. It is important to discover situational intentions and proper actions by deliberating on temporal abstractions to solve problems effectively. To understand the intention separated from changing task dynamics, we extend an empowerment-based regularization technique to situations with multiple tasks based on the framework of a generative adversarial network. Under the multitask environments with unknown dynamics, we focus on learning a reward and policy from the unlabeled expert examples. In this study, we define situational empowerment as the maximum of mutual information representing how an action conditioned on both a certain state and sub-task affects the future. Our proposed method derives the variational lower bound of the situational mutual information to optimize it. We simultaneously learn the transferable multi-task reward function and policy by adding an induced term to the objective function. By doing so, the multi-task reward function helps to learn a robust policy for environmental change. We validate the advantages of our approach on multi-task learning and multitask transfer learning. We demonstrate our proposed method has the robustness of both randomness and changing task dynamics. Finally, we prove that our method has significantly better performance and data efficiency than existing imitation learning methods on various benchmarks.
引用
收藏
页数:7
相关论文
共 50 条
  • [31] Multi-task Supervised Learning via Cross-learning
    Cervino, Juan
    Andres Bazerque, Juan
    Calvo-Fullana, Miguel
    Ribeiro, Alejandro
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 1381 - 1385
  • [32] Fine-Grained Driving Behavior Prediction via Context-Aware Multi-Task Inverse Reinforcement Learning
    Nishi, Kentaro
    Shimosaka, Masamichi
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 2281 - 2287
  • [33] Skills Regularized Task Decomposition for Multi-task Offline Reinforcement Learning
    Yoo, Minjong
    Cho, Sangwoo
    Woo, Honguk
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [34] Multi-task Deep Reinforcement Learning for Scalable Parallel Task Scheduling
    Zhang, Lingxin
    Qi, Qi
    Wang, Jingyu
    Sun, Haifeng
    Liao, Jianxin
    2019 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2019, : 2992 - 3001
  • [35] Learning task-relevant representations via rewards and real actions for reinforcement learning
    Yuan, Linghui
    Lu, Xiaowei
    Liu, Yunlong
    KNOWLEDGE-BASED SYSTEMS, 2024, 294
  • [36] A Hybrid Multi-Task Learning Approach for Optimizing Deep Reinforcement Learning Agents
    Varghese, Nelson Vithayathil
    Mahmoud, Qusay H.
    IEEE ACCESS, 2021, 9 : 44681 - 44703
  • [37] Variational Multi-Task Learning with Gumbel-Softmax Priors
    Shen, Jiayi
    Zhen, Xiantong
    Worring, Marcel
    Shao, Ling
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [38] Multi-task Scheduling of Multiple Agricultural Machinery via Reinforcement Learning and Genetic Algorithm
    Li, Lihang
    Jia, Liruizhi
    Liu, Shengquan
    Kong, Bo
    Liu, Yuan
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT I, ICIC 2024, 2024, 14862 : 70 - 81
  • [39] Learning to Branch for Multi-Task Learning
    Guo, Pengsheng
    Lee, Chen-Yu
    Ulbricht, Daniel
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [40] Learning to Branch for Multi-Task Learning
    Guo, Pengsheng
    Lee, Chen-Yu
    Ulbricht, Daniel
    25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,