Deep reinforcement learning with robust augmented reward sequence prediction for improving GNSS positioning

被引:0
|
作者
Tang, Jianhao [1 ,2 ]
Li, Zhenni [1 ,3 ]
Yu, Qingsong [4 ]
Zhao, Haoli [5 ]
Zeng, Kungan [6 ]
Zhong, Shiguang [4 ]
Wang, Qianming [1 ,7 ]
Xie, Kan [1 ,2 ]
Kuzin, Victor [8 ]
Xie, Shengli [2 ,9 ]
机构
[1] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[2] Guangdong Key Lab IoT Informat Technol, Guangzhou 510006, Peoples R China
[3] Guangdong HongKong Macao Joint Lab Smart Discrete, Guangzhou 510006, Guangdong, Peoples R China
[4] Guangzhou Haige Commun Grp Inc Co, Guangzhou 510006, Peoples R China
[5] Hebei Univ Technol, Sch Artificial Intelligence, Tianjin 300401, Peoples R China
[6] Sun Yat Sen Univ, Sch Software Engn, Zhuhai 519000, Peoples R China
[7] Taidou Microelect Technol Co Ltd, Guangzhou 510006, Peoples R China
[8] Acad Russian Engn Acad, Moscow, Russia
[9] Key Lab Intelligent Detect & Internet Things Mfg G, Guangzhou 510006, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep reinforcement learning; Robust augmented reward sequence prediction; Generalization; GNSS positioning;
D O I
10.1007/s10291-025-01824-w
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Data-driven technologies have shown promising potential for improving GNSS positioning, which can analyze observation data to learn the complex hidden characteristics of system models, without rigorous prior assumptions. However, in complex urban areas, the input observation data contain task-irrelevant noisy GNSS measurements arising from stochastic noise, such as signal reflections from tall buildings. Moreover, the problem of data distribution shift between the training and testing phases exists for dynamically changing environments. These problems limit the robustness and generalizability of the data-driven GNSS positioning methods in urban areas. In this paper, a novel deep reinforcement learning (DRL) method is proposed to improve the robustness and generalizability of the data-driven GNSS positioning. Specifically, to address the data distribution shift in dynamically changing environments, the robust Bellman operator (RBO) is employed into the DRL optimization to model the deviations in the data distribution and to enhance generalizability. To improve robustness against task-irrelevant noisy GNSS measurements, the long-term reward sequence prediction (LRSP) is adopted to learn robust representations by extracting task-relevant information from GNSS observations. Therefore, we develop a DRL method with robust augmented reward sequence prediction to correct the rough position solved by model-based methods. Moreover, a novel real-world GNSS positioning dataset is built, containing different scenes in urban areas. Our experiments were conducted on the public dataset Google smartphone decimeter challenge 2022 (GSDC2022) and the built dataset Guangzhou GNSS version 2 (GZGNSS-V2), which demonstrated that the proposed method can outperform model-based and state-of-the-art data-driven methods in terms of generalizability across different environments.
引用
收藏
页数:28
相关论文
共 50 条
  • [21] Deep Reinforcement Learning for Stock Prediction
    Zhang, Junhao
    Lei, Yifei
    SCIENTIFIC PROGRAMMING, 2022, 2022
  • [22] Imagination-Augmented Agents for Deep Reinforcement Learning
    Racaniere, Sebastien
    Weber, Theophane
    Reichert, David P.
    Buesing, Lars
    Guez, Arthur
    Rezende, Danilo
    Badia, Adria Puigdomenech
    Vinyals, Oriol
    Heess, Nicolas
    Li, Yujia
    Pascanu, Razvan
    Battaglia, Peter
    Hassabis, Demis
    Silver, David
    Wierstra, Daan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [23] Multitask Augmented Random Search in deep reinforcement learning
    Thanh, Le Tien
    Thang, Ta Bao
    Van Cuong, Le
    Binh, Huynh Thi Thanh
    APPLIED SOFT COMPUTING, 2024, 160
  • [24] Adversary Agnostic Robust Deep Reinforcement Learning
    Qu, Xinghua
    Gupta, Abhishek
    Ong, Yew-Soon
    Sun, Zhu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (09) : 6146 - 6157
  • [25] Robust Deep Reinforcement Learning for Quadcopter Control
    Deshpande, Aditya M.
    Minai, Ali A.
    Kumar, Manish
    IFAC PAPERSONLINE, 2021, 54 (20): : 90 - 95
  • [26] Doubly Robust Augmented Transfer for Meta-Reinforcement Learning
    Jiang, Yuankun
    Kan, Nuowen
    Li, Chenglin
    Dai, Wenrui
    Zou, Junni
    Xiong, Hongkai
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [27] AlphaSeq: Sequence Discovery With Deep Reinforcement Learning
    Shao, Yulin
    Liew, Soung Chang
    Wang, Taotao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (09) : 3319 - 3333
  • [28] Sequence Prediction with Unlabeled Data by Reward Function Learning
    Wu, Lijun
    Zhao, Li
    Qin, Tao
    Lai, Jianhuang
    Liu, Tie-Yan
    PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 3098 - 3104
  • [29] Improving Deep Reinforcement Learning With Mirror Loss
    Zhao, Jian
    Shu, Weide
    Zhao, Youpeng
    Zhou, Wengang
    Li, Houqiang
    IEEE TRANSACTIONS ON GAMES, 2023, 15 (03) : 337 - 347
  • [30] Improving Deep Reinforcement Learning with Knowledge Transfer
    Glatt, Ruben
    Reali Costa, Anna Helena
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 5036 - 5037