A lagrangian propagator for artificial neural networks in constraint programming

被引:11
|
作者
Lombardi, Michele [1 ]
Gualandi, Stefano [2 ]
机构
[1] Univ Bologna, Viale Risorgimento 2, I-40136 Bologna, Italy
[2] AntOptima SA, Via Aprica 26, CH-6900 Lugano, Switzerland
关键词
Constraint programming; Lagrangian relaxation; Neural networks; ALGORITHM; BOUNDS;
D O I
10.1007/s10601-015-9234-6
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper discusses a new method to perform propagation over a (two-layer, feed-forward) Neural Network embedded in a Constraint Programming model. The method is meant to be employed in Empirical Model Learning, a technique designed to enable optimal decision making over systems that cannot be modeled via conventional declarative means. The key step in Empirical Model Learning is to embed a Machine Learning model into a combinatorial model. It has been showed that Neural Networks can be embedded in a Constraint Programming model by simply encoding each neuron as a global constraint, which is then propagated individually. Unfortunately, this decomposition approach may lead to weak bounds. To overcome such limitation, we propose a new network-level propagator based on a non-linear Lagrangian relaxation that is solved with a subgradient algorithm. The method proved capable of dramatically reducing the search tree size on a thermal-aware dispatching problem on multicore CPUs. The overhead for optimizing the Lagrangian multipliers is kept within a reasonable level via a few simple techniques. This paper is an extended version of [27], featuring an improved structure, a new filtering technique for the network inputs, a set of overhead reduction techniques, and a thorough experimentation.
引用
收藏
页码:435 / 462
页数:28
相关论文
共 50 条
  • [31] Differentiable programming for online training of a neural artificial viscosity function within a staggered grid Lagrangian hydrodynamics scheme
    Melland, Pake
    Albright, Jason
    Urban, Nathan M.
    [J]. MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2021, 2 (02):
  • [32] Artificial neural networks
    Piuri, V
    Alippi, C
    [J]. JOURNAL OF SYSTEMS ARCHITECTURE, 1998, 44 (08) : 565 - 567
  • [33] Artificial Neural Networks
    Andrijic, Z. Ujevic
    [J]. KEMIJA U INDUSTRIJI-JOURNAL OF CHEMISTS AND CHEMICAL ENGINEERS, 2019, 68 (5-6): : 219 - 220
  • [34] Artificial neural networks
    Partridge, D
    Rae, S
    Wang, WJ
    [J]. JOURNAL OF THE ROYAL SOCIETY OF MEDICINE, 1999, 92 (07) : 385 - 385
  • [35] ARTIFICIAL NEURAL NETWORKS
    IVALL, T
    [J]. ELECTRONICS WORLD & WIRELESS WORLD, 1990, 96 (1649): : 191 - 193
  • [36] ARTIFICIAL NEURAL NETWORKS
    STRINGA, L
    DAPOR, M
    [J]. AEI AUTOMAZIONE ENERGIA INFORMAZIONE, 1994, 81 (03): : 325 - 331
  • [37] ARTIFICIAL NEURAL NETWORKS
    FULCHER, J
    [J]. COMPUTER STANDARDS & INTERFACES, 1994, 16 (03) : 183 - 184
  • [38] ARTIFICIAL NEURAL NETWORKS
    HOPFIELD, JJ
    [J]. IEEE CIRCUITS AND DEVICES MAGAZINE, 1988, 4 (05): : 3 - 10
  • [39] ARTIFICIAL NEURAL NETWORKS
    MAKHOUL, J
    [J]. INVESTIGATIVE RADIOLOGY, 1990, 25 (06) : 748 - 750
  • [40] Artificial neural networks
    Drew, PJ
    Monson, JRT
    [J]. SURGERY, 2000, 127 (01) : 3 - 11