Multilayer neural networks:: an experimental evaluation of on-line training methods

被引:21
|
作者
Martí, R [1 ]
El-Fallahi, A [1 ]
机构
[1] Univ Valencia, Dept Estadist & Invest Operat, E-46100 Valencia, Spain
关键词
D O I
10.1016/S0305-0548(03)00104-7
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Artificial neural networks (ANN) are inspired by the structure of biological neural networks and their ability to integrate knowledge and learning. In ANN training, the objective is to minimize the error over the training set. The most popular method for training these networks is back propagation, a gradient descent technique. Other non-linear optimization methods such as conjugate directions set or conjugate gradient have also been used for this purpose. Recently, metaheuristics such as simulated annealing, genetic algorithms or tabu search have been also adapted to this context. There are situations in which the necessary training data are being generated in real time and an extensive training is not possible. This "on-line" training arises in the context of optimizing a simulation. This paper presents extensive computational experiments to compare 12 "on-line" training methods over a collection of 45 functions from the literature within a short-term horizon. We propose a new method based on the tabu search methodology, which can compete in quality with the best previous approaches. Scope and purpose Artificial neural networks present a new paradigm for decision support that integrates knowledge and learning. They are inspired by biological neural systems where the nodes of the network represent the neurons and the arcs, the axons and dendrites. In recent years, there has been an increasing interest in ANN since they had proven very effectively in different contexts. In this paper we will focus on the prediction/estimation problem for a given function, where the input of the net is given by the values of the function variables and the output is the estimation of the function image. Specifically, we will consider the optimization problem that arises when training the net in the context of optimizing simulations (i.e. when the training time is limited). As far as we know, partial studies have been published, where a few training methods are compared over a limited set of instances. In this paper we present extensive computational experimentation of 12 different optimization methods over a set of 45 well-known functions. (C) 2003 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1491 / 1513
页数:23
相关论文
共 50 条
  • [1] A training data selection in on-line training for multilayer neural networks
    Hara, K
    Nakayama, K
    Kharaf, AAM
    [J]. IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, 1998, : 2247 - 2252
  • [2] Selection of minimum training data for generalization and on-line training by multilayer neural networks
    Hara, K
    Nakayama, K
    [J]. ICNN - 1996 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS. 1-4, 1996, : 436 - 441
  • [3] On-line learning from restricted training sets in multilayer neural networks
    Coolen, ACC
    Saad, D
    Xiong, YS
    [J]. EUROPHYSICS LETTERS, 2000, 51 (06): : 691 - 697
  • [4] Dynamic multilayer neural networks for nonlinear system on-line identification
    Yu, W
    Poznyak, AS
    Sanchez, EN
    [J]. PROCEEDINGS OF THE 2000 IEEE INTERNATIONAL SYMPOSIUM ON INTELLIGENT CONTROL, 2000, : 25 - 30
  • [5] Dynamics of on-line gradient descent learning for multilayer neural networks
    Saad, D
    Solla, SA
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 8: PROCEEDINGS OF THE 1995 CONFERENCE, 1996, 8 : 302 - 308
  • [6] Globally optimal parameters for on-line learning in multilayer neural networks
    Saad, D
    Rattray, M
    [J]. PHYSICAL REVIEW LETTERS, 1997, 79 (13) : 2578 - 2581
  • [7] A novel method for on-line training of dynamic neural networks
    Chowdhury, FN
    [J]. PROCEEDINGS OF THE 2001 IEEE INTERNATIONAL CONFERENCE ON CONTROL APPLICATIONS (CCA'01), 2001, : 161 - 166
  • [8] Adaptive stepsize algorithms for on-line training of neural networks
    Magoulas, GD
    Plagianakos, VP
    Vrahatis, MN
    [J]. NONLINEAR ANALYSIS-THEORY METHODS & APPLICATIONS, 2001, 47 (05) : 3425 - 3430
  • [9] Matrix-update for accelerated on-line learning in multilayer neural networks
    Bos, S
    [J]. JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1998, 31 (22): : L413 - L417
  • [10] Finite-size effects in on-line learning of multilayer neural networks
    Barber, D
    Saad, D
    Sollich, P
    [J]. EUROPHYSICS LETTERS, 1996, 34 (02): : 151 - 156