Subgradient-based neural networks for nonsmooth convex optimization problems

被引:96
|
作者
Xue, Xiaoping [1 ]
Bian, Wei [1 ]
机构
[1] Harbin Inst Technol, Dept Math, Harbin 150001, Peoples R China
关键词
convergence in finite time; differential inclusion; feasible region; neural network; nonsmooth convex optimization problem;
D O I
10.1109/TCSI.2008.920131
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper develops a neural network for solving the general nonsmooth convex optimization problems. The proposed neural network is modeled by a differential inclusion. Compared with the existing neural networks for solving nonsmooth convex optimization problems, this neural network has a wider domain for implementation. Under a suitable assumption on the constraint set, it is proved that for a given nonsmooth convex optimization problem and sufficiently large penalty parameters, any trajectory of the neural network can reach the feasible region in finite time and stays there thereafter. Moreover, we can prove that the trajectory of the neural network constructed by a differential inclusion and with arbitrarily given initial value, converges to the set consisting of the equilibrium points of the neural network, whose elements are all the optimal solutions of the primal constrained optimization problem. In particular, we give the condition that the equilibrium point set of the neural network coincides with the optimal solution set of the primal constrained optimization problem and the condition ensuring convergence to the optimal solution set in finite time. Furthermore, illustrative examples show the correctness of the results in this paper, and the good performance of the proposed neural network.
引用
收藏
页码:2378 / 2391
页数:14
相关论文
共 50 条