Soft-Clipping Swish: A Novel Activation Function for Deep Learning

被引:4
|
作者
Mercioni, Marina Adriana [1 ]
Holban, Stefan [1 ]
机构
[1] Politehn Univ Timisoara, Dept Comp Sci, Timisoara, Romania
关键词
activation function; classification; clipping; Computer Vision; Deep Learning; LeNet; ReLU; residual; ResNet; Sigmoid; soft; Swish;
D O I
10.1109/SACI51354.2021.9465622
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This study aims to contribute to the improvement of the network's performance through developing a novel activation function. Over time, many activation functions have been proposed in order to sole the issues of the previous functions. We note here more than 50 activation functions that have been proposed, some of them being very popular such as sigmoid, Rectified Linear Unit (ReLU), Swish, Mish but not only. The main idea of this study that stays behind our proposal is a simple one, based on a very popular function called Swish, which is a composition function, having in its componence sigmoid function and ReLU function. Starting from this activation function we decided to ignore the negative region in the w as the Rectified Linear Unit does but being different than that one mentioned through a nonlinear curse assured by the Swish positive region. The idea has been come up from a current function called Soft Clipping. We tested this proposal on more datasets in Computer Vision on classification tasks showing its high potential, here we mention MNIST, Fashion-MNIST, CIFAR-10, CIFAR-100 using two popular architectures: LeNet-5 and ResNet20 version 1.
引用
收藏
页码:225 / 230
页数:6
相关论文
共 50 条
  • [1] Soft Clipping Mish - A Novel Activation Function for Deep Learning
    Mercioni, Marina Adriana
    Holban, Stefan
    [J]. 2021 4TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMPUTER TECHNOLOGIES (ICICT 2021), 2021, : 13 - 17
  • [2] P-Swish: Activation Function with Learnable Parameters Based on Swish Activation Function in Deep Learning
    Mercioni, Marina Adriana
    Holban, Stefan
    [J]. 2020 14TH INTERNATIONAL SYMPOSIUM ON ELECTRONICS AND TELECOMMUNICATIONS (ISETC), 2020, : 40 - 43
  • [3] ALIASING REDUCTION IN SOFT-CLIPPING ALGORITHMS
    Esqueda, Fabian
    Valimaki, Vesa
    Bilbao, Stefan
    [J]. 2015 23RD EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2015, : 2014 - 2018
  • [4] Soft-clipping INGARCH models for time series of bounded counts
    Weiss, Christian H. H.
    Jahn, Malte
    [J]. STATISTICAL MODELLING, 2024, 24 (04) : 295 - 319
  • [5] PARAMETRIC FLATTEN-T SWISH: AN ADAPTIVE NON-LINEAR ACTIVATION FUNCTION FOR DEEP LEARNING
    Chieng, Hock Hung
    Wahid, Noorhaniza
    Ong, Pauline
    [J]. JOURNAL OF INFORMATION AND COMMUNICATION TECHNOLOGY-MALAYSIA, 2021, 20 (01): : 21 - 39
  • [6] Is it Time to Swish? Comparing Deep Learning Activation Functions Across NLP tasks
    Eger, Steffen
    Youssef, Paul
    Gurevych, Iryna
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4415 - 4424
  • [7] Smish: A Novel Activation Function for Deep Learning Methods
    Wang, Xueliang
    Ren, Honge
    Wang, Achuan
    [J]. ELECTRONICS, 2022, 11 (04)
  • [8] Enhancement of Deep Learning in Image Classification Performance Using Xception with the Swish Activation Function for Colorectal Polyp Preliminary Screening
    Jinsakul, Natinai
    Tsai, Cheng-Fa
    Tsai, Chia-En
    Wu, Pensee
    [J]. MATHEMATICS, 2019, 7 (12)
  • [9] A New Soft-Clipping Discrete Beta GARCH Model and Its Application on Measles Infection
    Chen, Huaping
    [J]. STATS, 2023, 6 (01): : 293 - 311
  • [10] A Novel Soft Margin Loss Function for Deep Discriminative Embedding Learning
    Yang, Zhao
    Liu, Tie
    Liu, Jiehao
    Wang, Li
    Zhao, Sai
    [J]. IEEE ACCESS, 2020, 8 : 202785 - 202794