Nonsmooth Optimization-Based Hyperparameter-Free Neural Networks for Large-Scale Regression

被引:0
|
作者
Karmitsa, Napsu [1 ]
Taheri, Sona [2 ]
Joki, Kaisa [3 ]
Paasivirta, Pauliina [4 ]
Bagirov, Adil M. [5 ]
Makela, Marko M. [3 ]
机构
[1] Univ Turku, Dept Comp, FI-20014 Turku, Finland
[2] RMIT Univ, Sch Sci, Melbourne 3000, Australia
[3] Univ Turku, Dept Math & Stat, FI-20014 Turku, Finland
[4] Siili Solut Oyj, FI-60100 Seinajoki, Finland
[5] Federat Univ Australia, Ctr Smart Analyt, Ballarat 3350, Australia
基金
澳大利亚研究理事会; 芬兰科学院;
关键词
machine learning; regression analysis; neural networks; L1-loss function; nonsmooth optimization; PERFORMANCE; REPRESENTATIONS; PARAMETERS; MACHINE;
D O I
10.3390/a16090444
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, a new nonsmooth optimization-based algorithm for solving large-scale regression problems is introduced. The regression problem is modeled as fully-connected feedforward neural networks with one hidden layer, piecewise linear activation, and the L1-loss functions. A modified version of the limited memory bundle method is applied to minimize this nonsmooth objective. In addition, a novel constructive approach for automated determination of the proper number of hidden nodes is developed. Finally, large real-world data sets are used to evaluate the proposed algorithm and to compare it with some state-of-the-art neural network algorithms for regression. The results demonstrate the superiority of the proposed algorithm as a predictive tool in most data sets used in numerical experiments.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] A Survey of Large-Scale Graph Neural Networks
    Xiao G.-Q.
    Li X.-Q.
    Chen Y.-D.
    Tang Z.
    Jiang W.-J.
    Li K.-L.
    Jisuanji Xuebao/Chinese Journal of Computers, 2024, 47 (01): : 148 - 171
  • [32] Training-free hyperparameter optimization of neural networks for electronic structures in matter
    Fiedler, Lenz
    Hoffmann, Nils
    Mohammed, Parvez
    Popoola, Gabriel A.
    Yovell, Tamar
    Oles, Vladyslav
    Ellis, J. Austin
    Rajamanickam, Sivasankaran
    Cangi, Attila
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2022, 3 (04):
  • [33] A Hypercube-Based Encoding for Evolving Large-Scale Neural Networks
    Stanley, Kenneth O.
    D'Ambrosio, David B.
    Gauci, Jason
    ARTIFICIAL LIFE, 2009, 15 (02) : 185 - 212
  • [34] Visual Odometry Based on Convolutional Neural Networks for Large-Scale Scenes
    Meng, Xuyang
    Fan, Chunxiao
    Ming, Yue
    TENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING (ICGIP 2018), 2019, 11069
  • [35] Detection of redundant traffic in large-scale communication networks based on logistic regression
    Wen X.
    Huang L.
    Zheng Y.
    Zhao H.
    International Journal of Reasoning-based Intelligent Systems, 2024, 16 (01) : 8 - 15
  • [36] Configuration and Optimization for Virtualization based Large-scale IP Networks Emulation
    Li Dawei
    2014 INTERNATIONAL CONFERENCE ON CYBER-ENABLED DISTRIBUTED COMPUTING AND KNOWLEDGE DISCOVERY (CYBERC), 2014, : 277 - 281
  • [37] Efficient Simulation-Based Toll Optimization for Large-Scale Networks
    Osorio, Carolina
    Atasoy, Bilge
    TRANSPORTATION SCIENCE, 2021, 55 (05) : 1010 - 1024
  • [38] Overlapping community detection based on conductance optimization in large-scale networks
    Gao, Yang
    Zhang, Hongli
    Zhang, Yue
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2019, 522 (69-79) : 69 - 79
  • [39] A SUPERCONDUCTING NEURAL CELL SUITABLE FOR LARGE-SCALE NEURAL NETWORKS
    HIDAKA, M
    AKERS, LA
    APPLIED SUPERCONDUCTIVITY, 1993, 1 (10-12) : 1907 - 1919
  • [40] A Population-Based Hybrid Approach for Hyperparameter Optimization of Neural Networks
    Japa, Luis
    Serqueira, Marcello
    Mendonca, Israel
    Aritsugi, Masayoshi
    Bezerra, Eduardo
    Gonzalez, Pedro Henrique
    IEEE ACCESS, 2023, 11 : 50752 - 50768