Regularization of deep neural network using a multisample memory model

被引:0
|
作者
Muhammad Tanveer [1 ]
Mohammad Yakoob Siyal [1 ]
Sheikh Faisal Rashid [2 ]
机构
[1] Nanyang Technological University,School of Electrical and Electronics Engineering
[2] Berlin Educational Technology Lab (EdTec),German Research Center for Artificial Intelligence (DFKI)
关键词
Deeper architecture; Overfitting; Regularization; Bag sampling; Memory model; Superfast convergence;
D O I
10.1007/s00521-024-10474-x
中图分类号
学科分类号
摘要
Deep convolutional neural networks (CNNs) are widely used in computer vision and have achieved significant performance for image classification tasks. Overfitting is a general problem in deep learning models that inhibit the generalization capability of deep models due to the presence of noise, the limited size of the training data, the complexity of the classifier, and the larger number of hyperparameters involved during training. Several techniques have been developed for overfitting inhibition, but in this research we focus only on regularization techniques. We propose a memory-based regularization technique to inhibit overfitting problems and generalize the performance of deep neural networks. Our backbone architectures receive input samples in bags rather than directly in batches to generate deep features. The proposed model receives input samples as queries and feeds them to the MAM (memory access module), which searches for the relevant items in memory and computes memory loss using Euclidean similarity measures. Our memory loss function incorporates intra-class compactness and inter-class separability at the feature level. Most surprisingly, the convergence rate of the proposed model is superfast, requiring only a few epochs to train both shallow and deeper models. In this study, we evaluate the performance of the memory model across several state-of-the-art (SOTA) deep learning architectures, including ReseNet18, ResNet50, ResNet101, VGG-16, AlexNet, and MobileNet, using the CIFAR-10 and CIFAR-100 datasets. The results show that the efficient memory model we have developed significantly outperforms almost all existing SOTA benchmarks by a considerable margin.
引用
收藏
页码:23295 / 23307
页数:12
相关论文
共 50 条
  • [1] Network as Regularization for Training Deep Neural Networks: Framework, Model and Performance
    Tian, Kai
    Xu, Yi
    Guan, Jihong
    Zhou, Shuigeng
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 6013 - 6020
  • [2] Batch Contrastive Regularization for Deep Neural Network
    Tanveer, Muhammad
    Tan, Hung Khoon
    Ng, Hui Fuang
    Leung, Maylor Karhang
    Chuah, Joon Huang
    PROCEEDINGS OF THE 12TH INTERNATIONAL JOINT CONFERENCE ON COMPUTATIONAL INTELLIGENCE (IJCCI), 2020, : 368 - 377
  • [3] A radial basis deep neural network process using the Bayesian regularization optimization for the monkeypox transmission model
    Akkilic, Ayse Nur
    Sabir, Zulqurnain
    Bhat, Shahid Ahmad
    Bulut, Hasan
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 235
  • [4] Implicit regularization of a deep augmented neural network model for human motion prediction
    Yadav, Gaurav Kumar
    Abdel-Nasser, Mohamed
    Rashwan, Hatem A.
    Puig, Domenec
    Nandi, G. C.
    APPLIED INTELLIGENCE, 2023, 53 (14) : 18027 - 18040
  • [5] Deep Convolutional Neural Network Regularization for Alcoholism Detection Using EEG Signals
    Mukhtar, Hamid
    Qaisar, Saeed Mian
    Zaguia, Atef
    SENSORS, 2021, 21 (16)
  • [6] Regularization of Deep Neural Network With Batch Contrastive Loss
    Tanveer, Muhammad
    Tan, Hung-Khoon
    Ng, Hui-Fuang
    Leung, Maylor Karhang
    Chuah, Joon Huang
    IEEE ACCESS, 2021, 9 : 124409 - 124418
  • [7] DEEP RECURRENT REGULARIZATION NEURAL NETWORK FOR SPEECH RECOGNITION
    Chien, Jen-Tzung
    Lu, Tsai-Wei
    2015 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING (ICASSP), 2015, : 4560 - 4564
  • [8] deep Neural Network Regularization (dNNR) on denoised Image
    Singh, Richa
    Dubey, Ashwani Kumar
    Kapoor, Rajiv
    International Journal of Intelligent Information Technologies, 2022, 18 (01)
  • [9] TIKHONOV REGULARIZATION FOR DEEP NEURAL NETWORK ACOUSTIC MODELING
    Chien, Jen-Tzung
    Lu, Tsai-Wei
    2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014, 2014, : 147 - 152
  • [10] Correction to: Implicit regularization of a deep augmented neural network model for human motion prediction
    Gaurav Kumar Yadav
    Mohamed Abdel-Nasser
    Hatem A. Rashwan
    Domenec Puig
    G. C. Nandi
    Applied Intelligence, 2025, 55 (6)