Temporal prediction model with context-aware data augmentation for robust visual reinforcement learning

被引:0
|
作者
Yue, Xinkai [1 ]
Ge, Hongwei [1 ]
He, Xin [1 ]
Hou, Yaqing [1 ]
机构
[1] College of Computer Science and Technology, Dalian University of Technology, Dalian, China
基金
中国国家自然科学基金;
关键词
Benchmarking - Forecasting - Learning systems - Pixels - Robotics;
D O I
10.1007/s00521-024-10251-w
中图分类号
学科分类号
摘要
While reinforcement learning has shown promising abilities to solve continuous control tasks from visual inputs, it remains a challenge to learn robust representations from high-dimensional observations and generalize to unseen environments with distracting elements. Recently, strong data augmentation has been applied to increase the diversity of the training data, but it may damage the task-relevant pixels and thus hinder the optimization of reinforcement learning. To this end, this paper proposes temporal prediction model with context-aware data augmentation (TPMC), a framework which incorporates context-aware strong augmentation into the dynamic model for learning robust policies. Specifically, TPMC utilizes the gradient-based saliency map to identify and preserve task-relevant pixels during strong augmentation, generating reliable augmented images for stable training. Moreover, the temporal prediction consistency between strong and weak augmented views is enforced to construct a contrastive objective for learning shared task-relevant representations. Extensive experiments are conducted to evaluate the performance on DMControl-GB benchmarks and several robotic manipulation tasks. Experimental results demonstrate that TPMC achieves superior data-efficiency and generalization to other state-of-the-art methods. © The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2024.
引用
收藏
页码:19337 / 19352
页数:15
相关论文
共 50 条
  • [41] Inland Vessel Travel Time Prediction via a Context-Aware Deep Learning Model
    Fan, Tengze
    Chen, Deshan
    Huang, Chen
    Tian, Chi
    Yan, Xinping
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2023, 11 (06)
  • [42] Enhancing CTR Prediction with Context-Aware Feature Representation Learning
    Wang, Fangye
    Wang, Yingxu
    Li, Dongsheng
    Gu, Hansu
    Lu, Tun
    Zhang, Peng
    Gu, Ning
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 343 - 352
  • [43] Spatial-Temporal Context-Aware Online Action Detection and Prediction
    Huang, Jingjia
    Li, Nannan
    Li, Thomas
    Liu, Shan
    Li, Ge
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (08) : 2650 - 2662
  • [44] Mobile Context-Aware Learning Model and Platform Research
    Wan, Benting
    Cai, Chao
    INTERNATIONAL CONFERENCE ON COMPLEX SCIENCE MANAGEMENT AND EDUCATION SCIENCE (CSMES 2013), 2013, : 427 - 432
  • [45] Uncertainty-Aware Data Augmentation for Offline Reinforcement Learning
    Su, Yunjie
    Kong, Yilun
    Wang, Xueqian
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [46] Context-Aware Mobile Learning
    Economides, Anastasios A.
    OPEN KNOWLEDGE SOCIETY: A COMPUTER SCIENCE AND INFORMATION SYSTEMS MANIFESTO, 2008, 19 : 213 - 220
  • [47] Unified Visual Perception Model for Context-aware Wearable AR
    Jang, Youngkyoon
    Woo, Woontack
    2013 IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY (ISMAR) - SCIENCE AND TECHNOLOGY, 2013,
  • [48] Data integration meta model based on context-aware
    Bai, Z., 1600, Advanced Institute of Convergence Information Technology, Myoungbo Bldg 3F,, Bumin-dong 1-ga, Seo-gu, Busan, 602-816, Korea, Republic of (06):
  • [49] Context-Aware weighted for visual tracking
    Ding, Xiaoxue
    Xu, Haixia
    Huang, Yunjia
    Liu, Yong
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 1598 - 1603
  • [50] Towards context-aware collaborative filtering by learning context-aware latent representations
    Liu, Xin
    Zhang, Jiyong
    Yan, Chenggang
    KNOWLEDGE-BASED SYSTEMS, 2020, 199