Contrastive Modules with Temporal Attention for Multi-Task Reinforcement Learning

被引:0
|
作者
Lan, Siming [1 ,2 ,3 ]
Zhang, Rui [2 ]
Yi, Qi [1 ,2 ,3 ]
Guo, Jiaming [2 ]
Peng, Shaohui [5 ]
Gao, Yunkai [1 ,2 ,3 ]
Wu, Fan [2 ,3 ,4 ,5 ]
Chen, Ruizhi [5 ]
Du, Zidong [2 ]
Hu, Xing [2 ,6 ]
Zhang, Xishan [2 ,3 ]
Li, Ling [4 ,5 ]
Chen, Yunji [2 ,4 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
[2] Chinese Acad Sci, State Key Lab Processors, Inst Comp Technol, Beijing, Peoples R China
[3] Cambricon Technol, Beijing, Peoples R China
[4] Univ Chinese Acad Sci, Beijing, Peoples R China
[5] Chinese Acad Sci, Inst Software, Intelligent Software Res Ctr, Beijing, Peoples R China
[6] Shanghai Innovat Ctr Processor Technol, SHIC, Shanghai, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the field of multi-task reinforcement learning, the modular principle, which involves specializing functionalities into different modules and combining them appropriately, has been widely adopted as a promising approach to prevent the negative transfer problem that performance degradation due to conflicts between tasks. However, most of the existing multi-task RL methods only combine shared modules at the task level, ignoring that there may be conflicts within the task. In addition, these methods do not take into account that without constraints, some modules may learn similar functions, resulting in restricting the model's expressiveness and generalization capability of modular methods. In this paper, we propose the Contrastive Modules with Temporal Attention(CMTA) method to address these limitations. CMTA constrains the modules to be different from each other by contrastive learning and combining shared modules at a finer granularity than the task level with temporal attention, alleviating the negative transfer within the task and improving the generalization ability and the performance for multi-task RL. We conducted the experiment on Meta-World, a multi-task RL benchmark containing various robotics manipulation tasks. Experimental results show that CMTA outperforms learning each task individually for the first time and achieves substantial performance improvements over the baselines. Our code can be found at https://github.com/niiceMing/CMTA.
引用
收藏
页数:17
相关论文
共 50 条
  • [1] Multi-Task Reinforcement Learning With Attention-Based Mixture of Experts
    Cheng, Guangran
    Dong, Lu
    Cai, Wenzhe
    Sun, Changyin
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (06) : 3811 - 3818
  • [2] Multi-task reinforcement learning in humans
    Momchil S. Tomov
    Eric Schulz
    Samuel J. Gershman
    [J]. Nature Human Behaviour, 2021, 5 : 764 - 773
  • [3] Multi-task reinforcement learning in humans
    Tomov, Momchil S.
    Schulz, Eric
    Gershman, Samuel J.
    [J]. NATURE HUMAN BEHAVIOUR, 2021, 5 (06) : 764 - +
  • [4] Sparse Multi-Task Reinforcement Learning
    Calandriello, Daniele
    Lazaric, Alessandro
    Restelli, Marcello
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [5] Multi-task Learning with Modular Reinforcement Learning
    Xue, Jianyong
    Alexandre, Frederic
    [J]. FROM ANIMALS TO ANIMATS 16, 2022, 13499 : 127 - 138
  • [6] Sparse multi-task reinforcement learning
    Calandriello, Daniele
    Lazaric, Alessandro
    Restelli, Marcello
    [J]. INTELLIGENZA ARTIFICIALE, 2015, 9 (01) : 5 - 20
  • [7] Episodic task agnostic contrastive training for multi-task learning?
    Zhou, Fan
    Chen, Yuyi
    Wen, Jun
    Zeng, Qiuhao
    Shui, Changjian
    Ling, Charles X.
    Yang, Shichun
    Wang, Boyu
    [J]. NEURAL NETWORKS, 2023, 162 : 34 - 45
  • [8] Unsupervised Task Clustering for Multi-task Reinforcement Learning
    Ackermann, Johannes
    Richter, Oliver
    Wattenhofer, Roger
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 222 - 237
  • [9] Service recommendation based on contrastive learning and multi-task learning
    Yu, Ting
    Zhang, Lihua
    Liu, Hailin
    Liu, Hongbing
    Wang, Jiaojiao
    [J]. COMPUTER COMMUNICATIONS, 2024, 213 : 285 - 295
  • [10] Multi-task Batch Reinforcement Learning with Metric Learning
    Li, Jiachen
    Quan Vuong
    Liu, Shuang
    Liu, Minghua
    Ciosek, Kamil
    Christensen, Henrik
    Su, Hao
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33