Kafnets: Kernel-based non-parametric activation functions for neural networks

被引:42
|
作者
Scardapane, Simone [1 ]
Van Vaerenbergh, Steven [3 ]
Totaro, Simone [2 ]
Uncini, Aurelio [1 ]
机构
[1] Sapienza Univ Rome, Dept Informat Engn Elect & Telecommun DIET, Via Eudossiana 18, I-00184 Rome, Italy
[2] Sapienza Univ Rome, Dept Stat Sci, Piazzale Aldo Moro 5, I-00185 Rome, Italy
[3] Univ Cantabria, Dept Commun Engn, Av Castros S-N, E-39005 Santander, Cantabria, Spain
关键词
Neural networks; Activation functions; Kernel methods; MULTILAYER FEEDFORWARD NETWORKS; TRAINABLE AMPLITUDE;
D O I
10.1016/j.neunet.2018.11.002
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks are generally built by interleaving (adaptable) linear layers with (fixed) nonlinear activation functions. To increase their flexibility, several authors have proposed methods for adapting the activation functions themselves, endowing them with varying degrees of flexibility. None of these approaches, however, have gained wide acceptance in practice, and research in this topic remains open. In this paper, we introduce a novel family of flexible activation functions that are based on an inexpensive kernel expansion at every neuron. Leveraging several properties of kernel-based models, we propose multiple variations for designing and initializing these kernel activation functions (KAFs), including a multidimensional scheme allowing to nonlinearly combine information from different paths in the network. The resulting KAFs can approximate any mapping defined over a subset of the real line, either convex or non-convex. Furthermore, they are smooth over their entire domain, linear in their parameters, and they can be regularized using any known scheme, including the use of l(1) penalties to enforce sparseness. To the best of our knowledge, no other known model satisfies all these properties simultaneously. In addition, we provide an overview on alternative techniques for adapting the activation functions, which is currently lacking in the literature. A large set of experiments validates our proposal. (C) 2018 Elsevier Ltd. All rights reserved.
引用
收藏
页码:19 / 32
页数:14
相关论文
共 50 条
  • [1] Non-parametric kernel-based estimation and simulation of precipitation amount
    Pavlides, Andrew
    Agou, Vasiliki D.
    Hristopulos, Dionissios T.
    [J]. JOURNAL OF HYDROLOGY, 2022, 612
  • [2] Recovery Guarantees for Kernel-based Clustering under Non-parametric Mixture Models
    Vankadara, Leena C.
    Bordt, Sebastian
    von Luxburg, Ulrike
    Ghoshdastidar, Debarghya
    [J]. 24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [3] Improving Graph Convolutional Networks with Non-Parametric Activation Functions
    Scardapane, Simone
    Van Vaerenbergh, Steven
    Comminiello, Danilo
    Uncini, Aurelio
    [J]. 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 872 - 876
  • [4] Approximate Kernel-Based Conditional Independence Tests for Fast Non-Parametric Causal Discovery
    Strobl, Eric, V
    Zhang, Kun
    Visweswaran, Shyam
    [J]. JOURNAL OF CAUSAL INFERENCE, 2019, 7 (01)
  • [5] Kernel-Based Non-parametric Clustering for Load Profiling of Big Smart Meter Data
    Pan, Erte
    Li, Husheng
    Song, Lingyang
    Han, Zhu
    [J]. 2015 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2015, : 2251 - 2255
  • [6] Kernel-based Generalized Cross-validation in Non-parametric Mixed-effect Models
    Xu, Wangli
    Zhu, Lixing
    [J]. SCANDINAVIAN JOURNAL OF STATISTICS, 2009, 36 (02) : 229 - 247
  • [7] KAF + RSigELU: a nonlinear and kernel-based activation function for deep neural networks
    Serhat Kiliçarslan
    Mete Celik
    [J]. Neural Computing and Applications, 2022, 34 : 13909 - 13923
  • [8] Learning Activation Functions by Means of Kernel Based Neural Networks
    Marra, Giuseppe
    Zanca, Dario
    Betti, Alessandro
    Gori, Marco
    [J]. ADVANCES IN ARTIFICIAL INTELLIGENCE, AI*IA 2019, 2019, 11946 : 418 - 430
  • [9] Robust localization based on non-parametric kernel technique
    Park, Chee-Hyun
    Chang, Joon-Hyuk
    [J]. ELECTRONICS LETTERS, 2022, 58 (22) : 850 - 852
  • [10] KAF plus RSigELU: a nonlinear and kernel-based activation function for deep neural networks
    Kilicarslan, Serhat
    Celik, Mete
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (16): : 13909 - 13923