Reinforcement learning guided Spearman dynamic opposite Gradient-based optimizer for numerical optimization and anchor clustering

被引:3
|
作者
Sun, Kangjian [1 ]
Huo, Ju [1 ]
Jia, Heming [2 ]
Yue, Lin [3 ]
机构
[1] Harbin Inst Technol, Sch Elect Engn & Automat, Harbin 150001, Peoples R China
[2] Sanming Univ, Sch Informat Engn, Sanming 365004, Peoples R China
[3] China Acad Railway Sci, Signal & Commun Res Inst, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Gradient-based optimizer; Reinforcement learning; Spearman rank correlation coefficient; Dynamic opposite; Numerical optimization; Anchor clustering; GLOBAL OPTIMIZATION; FEATURE-SELECTION; ALGORITHM;
D O I
10.1093/jcde/qwad109
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
As science and technology advance, the need for novel optimization techniques has led to an increase. The recently proposed metaheuristic algorithm, Gradient-based optimizer (GBO), is rooted in the gradient-based Newton's method. GBO has a more concrete theoretical foundation. However, gradient search rule (GSR) and local escaping operator (LEO) operators in GBO still have some shortcomings. The insufficient updating method and the simple selection process limit the search performance of the algorithm. In this paper, an improved version is proposed to compensate for the above shortcomings, called RL-SDOGBO. First, during the GSR phase, the Spearman rank correlation coefficient is used to determine weak solutions on which to perform dynamic opposite learning. This operation assists the algorithm to escape from local optima and enhance exploration capability. Secondly, to optimize the exploitation capability, reinforcement learning is used to guide the selection of solution update modes in the LEO operator. RL-SDOGBO is tested on 12 classical benchmark functions and 12 CEC2022 benchmark functions with seven representative metaheuristics, respectively. The impact of the improvements, the scalability and running time of the algorithm, and the balance of exploration and exploitation are analyzed and discussed. Combining the experimental results and some statistical results, RL-SDOGBO exhibits excellent numerical optimization performance and provides high-quality solutions in most cases. In addition, RL-SDOGBO is also used to solve the anchor clustering problem for small target detection, making it a more potential and competitive option. Graphical Abstract
引用
收藏
页码:12 / 33
页数:22
相关论文
共 50 条
  • [11] Gradient-based adaptive particle swarm optimizer with improved extremal optimization
    Xiaoli Zhao
    Jenq-Neng Hwang
    Zhijun Fang
    Guozhong Wang
    Applied Intelligence, 2018, 48 : 4646 - 4659
  • [12] Inverse Reinforcement Learning from a Gradient-based Learner
    Ramponi, Giorgia
    Drappo, Gianluca
    Restelli, Marcello
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [13] Gradient-based adaptive particle swarm optimizer with improved extremal optimization
    Zhao, Xiaoli
    Hwang, Jenq-Neng
    Fang, Zhijun
    Wang, Guozhong
    APPLIED INTELLIGENCE, 2018, 48 (12) : 4646 - 4659
  • [14] A gradient-based reinforcement learning approach to dynamic pricing in partially-observable environments
    Vengerov, David
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2008, 24 (07): : 687 - 693
  • [15] A dynamic neighborhood learning based particle swarm optimizer for global numerical optimization
    Nasir, Md
    Das, Swagatam
    Maity, Dipankar
    Sengupta, Soumyadip
    Halder, Udit
    Suganthan, P. N.
    INFORMATION SCIENCES, 2012, 209 : 16 - 36
  • [16] A Gradient-Based Reinforcement Learning Algorithm for Multiple Cooperative Agents
    Zhang, Zhen
    Wang, Dongqing
    Zhao, Dongbin
    Han, Qiaoni
    Song, Tingting
    IEEE ACCESS, 2018, 6 : 70223 - 70235
  • [17] Traffic Light Control with Policy Gradient-Based Reinforcement Learning
    Tas, Mehmet Bilge Han
    Ozkan, Kemal
    Saricicek, Inci
    Yazici, Ahmet
    32ND IEEE SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU 2024, 2024,
  • [18] Gradient-Based Inverse Risk-Sensitive Reinforcement Learning
    Mazumdar, Eric
    Ratliff, Lillian J.
    Fiez, Tanner
    Sastry, S. Shankar
    2017 IEEE 56TH ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2017,
  • [19] Leveraging Gradient-Based Optimizer and Deep Learning for Automated Soil Classification Model
    Alsolai, Hadeel
    Rizwanullah, Mohammed
    Maashi, Mashael
    Othman, Mahmoud
    Alneil, Amani A.
    Abdelmageed, Amgad Atta
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 76 (01): : 975 - 992
  • [20] Gradient-based Hyperparameter Optimization through Reversible Learning
    Maclaurin, Dougal
    Duvenaud, David
    Adams, Ryan P.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 2113 - 2122