Towards a Very Fast Feedforward Multilayer Neural Networks Training Algorithm

被引:6
|
作者
Bilski, Jaroslaw [1 ]
Kowalczyk, Bartosz [1 ]
Kisiel-Dorohinicki, Marek [2 ]
Siwocha, Agnieszka [3 ]
Zurada, Jacek [4 ]
机构
[1] Czestochowa Tech Univ, Dept Comp Engn, Al Armii Krajowej 36, PL-42200 Czestochowa, Poland
[2] AGH Univ Sci & Technol, Inst Comp Sci, PL-30059 Krakow, Poland
[3] Univ Social Sci, Informat Technol Inst, PL-90113 Lodz, Poland
[4] Univ Louisville, Dept Comp & Elect Engn, Louisville, KY 40292 USA
关键词
neural network training algorithm; QR decomposition; scaled Givens rotations; approximation; classification;
D O I
10.2478/jaiscr-2022-0012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
*This paper presents a novel fast algorithm for feedforward neural networks training. It is based on the Recursive Least Squares (RLS) method commonly used for designing adaptive filters. Besides, it utilizes two techniques of linear algebra, namely the orthogonal transformation method, called the Givens Rotations (GR), and the QR decomposition, creating the GQR (symbolically we write GR + QR = GQR) procedure for solving the normal equations in the weight update process. In this paper, a novel approach to the GQR algorithm is presented. The main idea revolves around reducing the computational cost of a single rotation by eliminating the square root calculation and reducing the number of multiplications. The proposed modification is based on the scaled version of the Givens rotations, denoted as SGQR. This modification is expected to bring a significant training time reduction comparing to the classic GQR algorithm. The paper begins with the introduction and the classic Givens rotation description. Then, the scaled rotation and its usage in the QR decomposition is discussed. The main section of the article presents the neural network training algorithm which utilizes scaled Givens rotations and QR decomposition in the weight update process. Next, the experiment results of the proposed algorithm are presented and discussed. The experiment utilizes several benchmarks combined with neural networks of various topologies. It is shown that the proposed algorithm outperforms several other commonly used methods, including well known Adam optimizer.
引用
收藏
页码:181 / 195
页数:15
相关论文
共 50 条
  • [1] NOVEL FAST TRAINING ALGORITHM FOR MULTILAYER FEEDFORWARD NEURAL NETWORK
    PARK, DJ
    JUN, BE
    KIM, JH
    [J]. ELECTRONICS LETTERS, 1992, 28 (06) : 543 - 545
  • [2] A NOVEL FAST FEEDFORWARD NEURAL NETWORKS TRAINING ALGORITHM
    Bilski, Jaroslaw
    Kowalczyk, Bartosz
    Marjanski, Andrzej
    Gandor, Michal
    Zurada, Jacek
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING RESEARCH, 2021, 11 (04) : 287 - 306
  • [3] A FAST NEW ALGORITHM FOR TRAINING FEEDFORWARD NEURAL NETWORKS
    SCALERO, RS
    TEPEDELENLIOGLU, N
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 1992, 40 (01) : 202 - 210
  • [4] A fast learning algorithm for training feedforward neural networks
    Goel, Ashok Kumar
    Saxena, Suresh C.
    Bhanot, Surekha
    [J]. INTERNATIONAL JOURNAL OF SYSTEMS SCIENCE, 2006, 37 (10) : 709 - 722
  • [5] A Second Order Training Algorithm for Multilayer Feedforward Neural Networks
    谭营
    何振亚
    邓超
    [J]. Journal of Southeast University(English Edition), 1997, (01) : 32 - 36
  • [6] A fast learning strategy for multilayer feedforward neural networks
    Chen, Huawei
    Zhong, Hualan
    Yuan, Haiying
    Jin, Fan
    [J]. WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 3019 - +
  • [7] Fast second order learning algorithm for feedforward multilayer neural networks and its applications
    Osowski, S
    Bojarczak, P
    Stodolski, M
    [J]. NEURAL NETWORKS, 1996, 9 (09) : 1583 - 1596
  • [8] An efficient recursive total least squares algorithm for training multilayer feedforward neural networks
    Choi, NJ
    Lim, JS
    Sung, KM
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2005, PT 1, PROCEEDINGS, 2005, 3496 : 558 - 565
  • [9] Exponentiated backpropagation algorithm for multilayer feedforward neural networks
    Srinivasan, N
    Ravichandran, V
    Chan, KL
    Vidhya, JR
    Ramakirishnan, S
    Krishnan, SM
    [J]. ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 327 - 331
  • [10] A hybrid training algorithm for feedforward neural networks
    Ben Nasr, Mounir
    Chtourou, Mohamed
    [J]. NEURAL PROCESSING LETTERS, 2006, 24 (02) : 107 - 117