Neural network for nonsmooth pseudoconvex optimization with general convex constraints

被引:53
|
作者
Bian, Wei [1 ,2 ]
Ma, Litao [1 ,3 ]
Qin, Sitian [4 ]
Xue, Xiaoping [1 ]
机构
[1] Harbin Inst Technol, Dept Math, Harbin 150001, Heilongjiang, Peoples R China
[2] Harbin Inst Technol, Inst Adv Study Math, Harbin 150001, Heilongjiang, Peoples R China
[3] Hebei Univ Engn, Sch Sci, Handan 056038, Peoples R China
[4] Harbin Inst Technol Weihai, Dept Math, Weihai 264209, Peoples R China
关键词
Neural network; Nonsmooth pseudoconvex optimization; Differential inclusion; Smoothing method; EXPONENTIAL STABILITY; MINIMIZATION; FINITE;
D O I
10.1016/j.neunet.2018.01.008
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a one-layer recurrent neural network is proposed for solving a class of nonsmooth, pseudoconvex optimization problems with general convex constraints. Based on the smoothing method, we construct a new regularization function, which does not depend on any information of the feasible region. Thanks to the special structure of the regularization function, we prove the global existence, uniqueness and "slow solution'' character of the state of the proposed neural network. Moreover, the state solution of the proposed network is proved to be convergent to the feasible region in finite time and to the optimal solution set of the related optimization problem subsequently. In particular, the convergence of the state to an exact optimal solution is also considered in this paper. Numerical examples with simulation results are given to show the efficiency and good characteristics of the proposed network. In addition, some preliminary theoretical analysis and application of the proposed network for a wider class of dynamic portfolio optimization are included. (c) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1 / 14
页数:14
相关论文
共 50 条