HOMEOSTATIC LEARNING RULE FOR ARTIFICIAL NEURAL NETWORKS

被引:0
|
作者
Ruzek, M. [1 ]
机构
[1] Czech Inst Informat Robot & Cybernet, Prague, Czech Republic
关键词
artificial neural network; learning rule; biological neuron;
D O I
10.14311/NNW.2018.28.011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article presents an improvement of learning algorithm for an artificial neural network that makes the learning process more similar to a biological neuron, but still simple enough to be easily programmed. This idea is based on autonomous artificial neurons that are working together and at same time competing for resources; every neuron is trying to be better than the others, but also needs the feed back from other neurons. The proposed artificial neuron has similar forward signal processing as the standard perceptron; the main difference is the learning phase. The learning process is based on observing the weights of other neurons, but only in biologically plausible way, no back propagation of error or 'teacher' is allowed. The neuron is sending the signal in a forward direction into the higher layer, while the information about its function is being propagated in the opposite direction. This information does not have the form of energy, it is the observation of how the neuron's output is accepted by the others. The neurons are trying to find such setting of their internal parameters that are optimal for the whole network. For this algorithm, it is necessary that the neurons are organized in layers. The tests proved the viability of this concept - the learning process is slower; but has other advantages, such as resistance against catastrophic interference or higher generalization.
引用
收藏
页码:179 / 189
页数:11
相关论文
共 50 条
  • [31] Learning flat representations with artificial neural networks
    Vlad Constantinescu
    Costin Chiru
    Tudor Boloni
    Adina Florea
    Robi Tacutu
    Applied Intelligence, 2021, 51 : 2456 - 2470
  • [32] Constrained-learning in Artificial Neural Networks
    Parra-Hernández, R
    2003 IEEE PACIFIC RIM CONFERENCE ON COMMUNICATIONS, COMPUTERS, AND SIGNAL PROCESSING, VOLS 1 AND 2, CONFERENCE PROCEEDINGS, 2003, : 352 - 355
  • [33] Advances in artificial neural networks and machine learning
    Prieto, Alberto
    Atencia, Miguel
    Sandoval, Francisco
    NEUROCOMPUTING, 2013, 121 : 1 - 4
  • [34] Evolutionary artificial neural networks for competitive learning
    Brown, AD
    Card, HC
    1997 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, 1997, : 1558 - 1562
  • [35] Statistical inference: learning in artificial neural networks
    Yang, HH
    Murata, N
    Amari, S
    TRENDS IN COGNITIVE SCIENCES, 1998, 2 (01) : 4 - 10
  • [36] Variational learning for quantum artificial neural networks
    Tacchino, Francesco
    Barkoutsos, Panagiotis Kl
    Macchiavello, Chiara
    Gerace, Dario
    Tavernelli, Ivano
    Bajoni, Daniele
    IEEE INTERNATIONAL CONFERENCE ON QUANTUM COMPUTING AND ENGINEERING (QCE20), 2020, : 130 - 136
  • [37] Learning flat representations with artificial neural networks
    Constantinescu, Vlad
    Chiru, Costin
    Boloni, Tudor
    Florea, Adina
    Tacutu, Robi
    APPLIED INTELLIGENCE, 2021, 51 (04) : 2456 - 2470
  • [38] ARTIFICIAL NEURAL NETWORKS LEARNING IN ROC SPACE
    Castro, Cristiano Leite
    Braga, Antonio Padua
    IJCCI 2009: PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON COMPUTATIONAL INTELLIGENCE, 2009, : 484 - +
  • [39] Framework for the interactive learning of artificial neural networks
    Uzak, Matus
    Jaksa, Rudolf
    ARTIFICIAL NEURAL NETWORKS - ICANN 2006, PT 1, 2006, 4131 : 103 - 112
  • [40] Overestimation Trap of Artificial Neural Network: Learning the Rule of PRBS
    Shu, Liang
    Li, Jianqiang
    Wan, Zhiquan
    Zhang, Wenjia
    Fu, Songnian
    Xu, Kun
    2018 EUROPEAN CONFERENCE ON OPTICAL COMMUNICATION (ECOC), 2018,