Filtering-based Layer-wise Parameter Update Method for Training a Neural Network

被引:0
|
作者
Ji, Siyu [1 ]
Zhai, Kaikai [1 ]
Wen, Chenglin [1 ]
机构
[1] Hangzhou Dianzi Univ, Inst Syst Sci & Control Engn, Hangzhou, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Neural Network; Parameter Training; Gradient Descent; Kalman Filtering; Extended Kalman Filtering;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Aiming at the difficulties in modeling variable nonlinear systems with noise interference, a network model with optimal generalization ability is established to realize the identification of the system. Traditional network parameter training method, like gradient descent method and least squares are centralized, and it is difficult to adaptively update the model parameters according to changes in the system. Firstly, in order to adaptively update the network parameters and quickly reflect the changes in the input and output of the system, the network weights are used as time-varying parameters, and some parameters in the network are updated by Kalman filtering algorithm. Then, in order to further improve the generalization ability of the network, EKF is used to update all the parameters in the network. Finally, the effectiveness of the algorithm is verified by an example of the standard data set UCI-ccpp.
引用
收藏
页码:389 / 394
页数:6
相关论文
共 50 条
  • [1] Layer-wise Pre-training Mechanism Based on Neural Network for Epilepsy Detection
    Lin, Zichao
    Gu, Zhenghui
    Li, Yinghao
    Yu, Zhuliang
    Li, Yuanqing
    [J]. 2020 12TH INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATIONAL INTELLIGENCE (ICACI), 2020, : 224 - 227
  • [2] Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training
    Zhou, Yefan
    Pang, Tianyu
    Liu, Keqin
    Martin, Charles H.
    Mahoney, Michael W.
    Yang, Yaoqing
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Layer-Wise Compressive Training for Convolutional Neural Networks
    Grimaldi, Matteo
    Tenace, Valerio
    Calimera, Andrea
    [J]. FUTURE INTERNET, 2019, 11 (01)
  • [4] A Layer-Wise Ensemble Technique for Binary Neural Network
    Xi, Jiazhen
    Yamauchi, Hiroyuki
    [J]. INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2021, 35 (08)
  • [5] The layer-wise method and the backpropagation hybrid approach to learning a feedforward neural network
    Rubanov, NS
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2000, 11 (02): : 295 - 305
  • [6] Post-training deep neural network pruning via layer-wise calibration
    Lazarevich, Ivan
    Kozlov, Alexander
    Malinin, Nikita
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2021), 2021, : 798 - 805
  • [7] Layer-Wise Training to Create Efficient Convolutional Neural Networks
    Zeng, Linghua
    Tian, Xinmei
    [J]. NEURAL INFORMATION PROCESSING (ICONIP 2017), PT II, 2017, 10635 : 631 - 641
  • [8] Layer-wise learning based stochastic gradient descent method for the optimization of deep convolutional neural network
    Zheng, Qinghe
    Tian, Xinyu
    Jiang, Nan
    Yang, Mingqiang
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 37 (04) : 5641 - 5654
  • [9] Network with Sub-networks: Layer-wise Detachable Neural Network
    Fuengfusin, Ninnart
    Tamukoh, Hakaru
    [J]. JOURNAL OF ROBOTICS NETWORKING AND ARTIFICIAL LIFE, 2021, 7 (04): : 240 - 244
  • [10] Craft Distillation: Layer-wise Convolutional Neural Network Distillation
    Blakeney, Cody
    Li, Xiaomin
    Yan, Yan
    Zong, Ziliang
    [J]. 2020 7TH IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND CLOUD COMPUTING (CSCLOUD 2020)/2020 6TH IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING AND SCALABLE CLOUD (EDGECOM 2020), 2020, : 252 - 257