Kernel Recursive Least-Squares Temporal Difference Algorithms with Sparsification and Regularization

被引:1
|
作者
Zhang, Chunyuan [1 ,2 ]
Zhu, Qingxin [1 ]
Niu, Xinzheng [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Engn & Comp Sci, Chengdu 611731, Peoples R China
[2] Hainan Univ, Coll Informat Sci & Technol, Haikou 570228, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1155/2016/2305854
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
By combining with sparse kernel methods, least-squares temporal difference (LSTD) algorithms can construct the feature dictionary automatically and obtain a better generalization ability. However, the previous kernel-based LSTD algorithms do not consider regularization and their sparsification processes are batch or offline, which hinder their widespread applications in online learning problems. In this paper, we combine the following five techniques and propose two novel kernel recursive LSTD algorithms: (i) online sparsification, which can cope with unknown state regions and be used for online learning, (ii) L-2 and L-1 regularization, which can avoid overfitting and eliminate the influence of noise, (iii) recursive least squares, which can eliminate matrix-inversion operations and reduce computational complexity, (iv) a sliding-window approach, which can avoid caching all history samples and reduce the computational cost, and (v) the fixed-point subiteration and online pruning, which can make L-1 regularization easy to implement. Finally, simulation results on two 50-state chain problems demonstrate the effectiveness of our algorithms.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Graph Kernel Recursive Least-Squares Algorithms
    Gogineni, Vinay Chakravarthi
    Naumova, Valeriya
    Werner, Stefan
    Huang, Yih-Fang
    [J]. 2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 2072 - 2076
  • [2] Recursive Least-Squares Temporal Difference With Gradient Correction
    Song, Tianheng
    Li, Dazi
    Yang, Weimin
    Hirasawa, Kotaro
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (08) : 4251 - 4264
  • [3] Multikernel Recursive Least-Squares Temporal Difference Learning
    Zhang, Chunyuan
    Zhu, Qingxin
    Niu, Xinzheng
    [J]. INTELLIGENT COMPUTING METHODOLOGIES, ICIC 2016, PT III, 2016, 9773 : 205 - 217
  • [4] The kernel recursive least-squares algorithm
    Engel, Y
    Mannor, S
    Meir, R
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2004, 52 (08) : 2275 - 2285
  • [5] Linear least-squares algorithms for temporal difference learning
    Bradtke, SJ
    Barto, AG
    [J]. MACHINE LEARNING, 1996, 22 (1-3) : 33 - 57
  • [6] Regularization for the Kernel Recursive Least Squares CMAC
    Laufer, C.
    Coghill, G.
    [J]. 2012 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2012,
  • [7] Deep kernel recursive least-squares algorithm
    Hossein Mohamadipanah
    Mahdi Heydari
    Girish Chowdhary
    [J]. Nonlinear Dynamics, 2021, 104 : 2515 - 2530
  • [8] Deep kernel recursive least-squares algorithm
    Mohamadipanah, Hossein
    Heydari, Mahdi
    Chowdhary, Girish
    [J]. NONLINEAR DYNAMICS, 2021, 104 (03) : 2515 - 2530
  • [9] On the Regularization of Recursive Least-Squares Adaptive Algorithms Using Line Search Methods
    Stanciu, Cristian-Lucian
    Anghel, Cristian
    Ficiu, Ionut-Dorinel
    Elisei-Iliescu, Camelia
    Udrea, Mihnea-Radu
    Stanciu, Lucian
    [J]. ELECTRONICS, 2024, 13 (08)
  • [10] Quaternion kernel recursive least-squares algorithm
    Wang, Gang
    Qiao, Jingci
    Xue, Rui
    Peng, Bei
    [J]. SIGNAL PROCESSING, 2021, 178