Triple-GAIL: A Multi-Modal Imitation Learning Framework with Generative Adversarial Nets

被引:0
|
作者
Fei, Cong [1 ,2 ]
Wang, Bin [1 ]
Zhuang, Yuzheng [1 ]
Zhang, Zongzhang [3 ]
Hao, Jianye [1 ]
Zhang, Hongbo [1 ]
Ji, Xuewu [2 ]
Liu, Wulong [1 ]
机构
[1] Huawei Noahs Ark Lab, Shenzhen, Peoples R China
[2] Tsinghua Univ, Beijing, Peoples R China
[3] Nanjing Univ, Nanjing, Jiangsu, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generative adversarial imitation learning (GAIL) has shown promising results by taking advantage of generative adversarial nets, especially in the field of robot learning. However, the requirement of isolated single modal demonstrations limits the scalability of the approach to real world scenarios such as autonomous vehicles' demand for a proper understanding of human drivers' behavior. In this paper, we propose a novel multi-modal GAIL framework, named Triple-GAIL, that is able to learn skill selection and imitation jointly from both expert demonstrations and continuously generated experiences with data augmentation purpose by introducing an auxiliary skill selector. We provide theoretical guarantees on the convergence to optima for both of the generator and the selector respectively. Experiments on real driver trajectories and real-time strategy game datasets demonstrate that Triple-GAIL can better fit multi-modal behaviors close to the demonstrators and outperforms state-of-the-art methods.
引用
收藏
页码:2929 / 2935
页数:7
相关论文
共 50 条
  • [21] Non-rigid multi-modal brain image registration based on two-stage generative adversarial nets
    Zhu, Xingxing
    Huang, Zhiwen
    Ding, Mingyue
    Zhang, Xuming
    NEUROCOMPUTING, 2022, 505 : 44 - 57
  • [22] Situated robot learning for multi-modal instruction and imitation of grasping
    Steil, M
    Röthling, F
    Haschke, R
    Ritter, H
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2004, 47 (2-3) : 129 - 141
  • [23] BAGAIL: Multi-modal imitation learning from imbalanced demonstrations
    Gu, Sijia
    Zhu, Fei
    NEURAL NETWORKS, 2024, 174
  • [24] A unified framework for multi-modal federated learning
    Xiong, Baochen
    Yang, Xiaoshan
    Qi, Fan
    Xu, Changsheng
    NEUROCOMPUTING, 2022, 480 : 110 - 118
  • [25] Multi-modal generative adversarial networks for traffic event detection in smart cities
    Chen, Qi
    Wang, Wei
    Huang, Kaizhu
    De, Suparna
    Coenen, Frans
    EXPERT SYSTEMS WITH APPLICATIONS, 2021, 177
  • [26] DBGAN: Dual Branch Generative Adversarial Network for Multi-Modal MRI Translation
    Lyu, Jun
    Yan, Shouang
    Hossain, M. Shamim
    ACM Transactions on Multimedia Computing, Communications and Applications, 2024, 20 (08)
  • [27] A Framework of Multi-modal Corpus for Mandarin Learning
    Liu, Yang
    Yang, Chunting
    2009 IITA INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS ENGINEERING, PROCEEDINGS, 2009, : 476 - 479
  • [28] DiamondGAN: Unified Multi-modal Generative Adversarial Networks for MRI Sequences Synthesis
    Li, Hongwei
    Paetzold, Johannes C.
    Sekuboyina, Anjany
    Kofler, Florian
    Zhang, Jianguo
    Kirschke, Jan S.
    Wiestler, Benedikt
    Menze, Bjoern
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT IV, 2019, 11767 : 795 - 803
  • [29] Free form deformation and symmetry constraint-based multi-modal brain image registration using generative adversarial nets
    Zhu, Xingxing
    Ding, Mingyue
    Zhang, Xuming
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2023, 8 (04) : 1492 - 1506
  • [30] Multi-modal False Information Detection Based on Adversarial Learning
    Tian, Tian
    Liu, Yudong
    Sun, Mengzhu
    Zhang, Xi
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,