Regularization of deep neural network using a multisample memory model

被引:0
|
作者
Muhammad Tanveer [1 ]
Mohammad Yakoob Siyal [1 ]
Sheikh Faisal Rashid [2 ]
机构
[1] Nanyang Technological University,School of Electrical and Electronics Engineering
[2] Berlin Educational Technology Lab (EdTec),German Research Center for Artificial Intelligence (DFKI)
关键词
Deeper architecture; Overfitting; Regularization; Bag sampling; Memory model; Superfast convergence;
D O I
10.1007/s00521-024-10474-x
中图分类号
学科分类号
摘要
Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the limited size of the training data, the complexity of the classifier, and the larger number of hyperparameters involved during training. Several techniques have been developed for overfitting inhibition, but in this research we focus only on regularization techniques. We propose a memory-based regularization technique to inhibit overfitting problems and generalize the performance of deep neural networks. Our backbone architectures receive input samples in bags rather than directly in batches to generate deep features. The proposed model receives input samples as queries and feeds them to the MAM (memory access module), which searches for the relevant items in memory and computes memory loss using Euclidean similarity measures. Our memory loss function incorporates intra-class compactness and inter-class separability at the feature level. Most surprisingly, the convergence rate of the proposed model is superfast, requiring only a few epochs to train both shallow and deeper models. In this study, we evaluate the performance of the memory model across several state-of-the-art (SOTA) deep learning architectures, including ReseNet18, ResNet50, ResNet101, VGG-16, AlexNet, and MobileNet, using the CIFAR-10 and CIFAR-100 datasets. The results show that the efficient memory model we have developed significantly outperforms almost all existing SOTA benchmarks by a considerable margin.
引用
收藏
页码:23295 / 23307
页数:12
相关论文
共 50 条
  • [21] Pulsar Candidate Recognition Using Deep Neural Network Model
    Yin, Qian
    Wang, Yan
    Zheng, Xin
    Zhang, Jikai
    ELECTRONICS, 2022, 11 (14)
  • [22] Haze Prediction Model Using Deep Recurrent Neural Network
    Shang, Kailin
    Chen, Ziyi
    Liu, Zhixin
    Song, Lihong
    Zheng, Wenfeng
    Yang, Bo
    Liu, Shan
    Yin, Lirong
    ATMOSPHERE, 2021, 12 (12)
  • [23] Air Quality Prediction Using a Deep Neural Network Model
    Cho, Kyunghak
    Lee, Byoung-Young
    Kwon, Myeongheum
    Kim, Seogcheol
    JOURNAL OF KOREAN SOCIETY FOR ATMOSPHERIC ENVIRONMENT, 2019, 35 (02) : 214 - 225
  • [24] A novel radial basis Bayesian regularization deep neural network for the Maxwell nanofluid applied on the Buongiorno model
    Sabir, Zulqurnain
    Akkurt, Nevzat
    Ben Said, Salem
    ARABIAN JOURNAL OF CHEMISTRY, 2023, 16 (06)
  • [25] MaxDropout: Deep Neural Network Regularization Based on Maximum Output Values
    Goncalves do Santos, Claudio Filipi
    Colombo, Danilo
    Roder, Mateus
    Papa, Joao Paulo
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2671 - 2676
  • [26] Deep Neural Network Regularization for Feature Selection in Learning-to-Rank
    Rahangdale, Ashwini
    Raut, Shital
    IEEE ACCESS, 2019, 7 : 53988 - 54006
  • [27] Comparison of Regularization Constraints in Deep Neural Network based Speaker Adaptation
    Shen, Peng
    Lu, Xugang
    Kawai, Hisashi
    2016 10TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2016,
  • [28] Group Pruning with Group Sparse Regularization for Deep Neural Network Compression
    Wu, Chenglu
    Pang, Wei
    Liu, Hao
    Lu, Shengli
    2019 IEEE 4TH INTERNATIONAL CONFERENCE ON SIGNAL AND IMAGE PROCESSING (ICSIP 2019), 2019, : 325 - 329
  • [29] A neural network model of spatial memory
    Pessa, E
    Montesanto, A
    Penna, MP
    ADVANCES IN INTELLIGENT SYSTEMS, 1997, 41 : 485 - 490
  • [30] Neural Network Model of Memory Retrieval
    Recanatesi, Stefano
    Katkov, Mikhail
    Romani, Sandro
    Tsodyks, Misha
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2015, 9