Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression

被引:0
|
作者
Karmitsa, Napsu [1 ]
Taheri, Sona [2 ]
Joki, Kaisa [3 ]
Paasivirta, Pauliina [4 ]
Bagirov, Adil M. [5 ]
Makela, Marko M. [3 ]
机构
[1] Univ Turku, Dept Comp, FI-20014 Turku, Finland
[2] RMIT Univ, Sch Sci, Melbourne 3000, Australia
[3] Univ Turku, Dept Math & Stat, FI-20014 Turku, Finland
[4] Siili Solut Oyj, FI-60100 Seinajoki, Finland
[5] Federat Univ Australia, Ctr Smart Analyt, Ballarat 3350, Australia
基金
澳大利亚研究理事会; 芬兰科学院;
关键词
machine learning; regression analysis; neural networks; L1-loss function; nonsmooth optimization; PERFORMANCE; REPRESENTATIONS; PARAMETERS; MACHINE;
D O I
10.3390/a16090444
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a new nonsmooth optimization-based algorithm for solving large-scale regression problems is introduced. The regression problem is modeled as fully-connected feedforward neural networks with one hidden layer, piecewise linear activation, and the L1-loss functions. A modified version of the limited memory bundle method is applied to minimize this nonsmooth objective. In addition, a novel constructive approach for automated determination of the proper number of hidden nodes is developed. Finally, large real-world data sets are used to evaluate the proposed algorithm and to compare it with some state-of-the-art neural network algorithms for regression. The results demonstrate the superiority of the proposed algorithm as a predictive tool in most data sets used in numerical experiments.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Hyperparameter optimization of neural networks based on Q-learning
    Xin Qi
    Bing Xu
    Signal, Image and Video Processing, 2023, 17 : 1669 - 1676
  • [22] Hyperparameter Tuning for Deep Neural Networks Based Optimization Algorithm
    Vidyabharathi, D.
    Mohanraj, V.
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 36 (03): : 2559 - 2573
  • [23] The Modified HZ Conjugate Gradient Algorithm for Large-Scale Nonsmooth Optimization
    Yuan, Gonglin
    Sheng, Zhou
    Liu, Wenjie
    PLOS ONE, 2016, 11 (10):
  • [24] New limited memory bundle method for large-scale nonsmooth optimization
    Haarala, M
    Miettinen, K
    Mäkelä, MM
    OPTIMIZATION METHODS & SOFTWARE, 2004, 19 (06): : 673 - 692
  • [25] Numerical collapse simulation of large-scale structural systems using an optimization-based algorithm
    Sivaselvan, Mettupalayam V.
    Lavan, Oren
    Dargush, Gary F.
    Kurino, Haruhiko
    Hyodo, Yo
    Fukuda, Ryusuke
    Sato, Kiochi
    Apostolakis, Georgios
    Reinhorn, Andre M.
    EARTHQUAKE ENGINEERING & STRUCTURAL DYNAMICS, 2009, 38 (05): : 655 - 677
  • [26] Resource Optimization-Based Software Risk Reduction Model for Large-Scale Application Development
    Shahzad, Basit
    Fazal-e-Amin
    Abro, Ahsanullah
    Imran, Muhammad
    Shoaib, Muhammad
    SUSTAINABILITY, 2021, 13 (05) : 1 - 17
  • [27] Marginalized Neural Network Mixtures for Large-Scale Regression
    Lazaro-Gredilla, Miguel
    Figueiras-Vidal, Anibal R.
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2010, 21 (08): : 1345 - 1351
  • [28] A particle swarm optimization-based approach to tackling simulation optimization of stochastic, large-scale and complex systems
    Lu, Ming
    Wu, Da-peng
    Zhang, Jian-ping
    ADVANCES IN MACHINE LEARNING AND CYBERNETICS, 2006, 3930 : 528 - 537
  • [29] On the Large-Scale Transferability of Convolutional Neural Networks
    Zheng, Liang
    Zhao, Yali
    Wang, Shengjin
    Wang, Jingdong
    Yang, Yi
    Tian, Qi
    TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING: PAKDD 2018 WORKSHOPS, 2018, 11154 : 27 - 39
  • [30] SNNSim: Investigation and Optimization of Large-Scale Analog Spiking Neural Networks Based on Flash Memory Devices
    Ko, Jong Hyun
    Kwon, Dongseok
    Hwang, Joon
    Lee, Kyu-Ho
    Oh, Seongbin
    Kim, Jeonghyun
    Im, Jiseong
    Koo, Ryun-Han
    Kim, Jae-Joon
    Lee, Jong-Ho
    ADVANCED INTELLIGENT SYSTEMS, 2024, 6 (04)