LMix: regularization strategy for convolutional neural networks

被引:4
|
作者
Yan, Linyu [1 ]
Zheng, Kunpeng [1 ]
Xia, Jinyao [1 ]
Li, Ke [1 ]
Ling, Hefei [2 ]
机构
[1] Hubei Univ Technol, Wuhan, Hubei, Peoples R China
[2] Huazhong Univ Sci & Technol, Wuhan, Hubei, Peoples R China
基金
中国国家自然科学基金;
关键词
Mixup; Data augmentation; Deep convolutional neural networks; Regularization strategies;
D O I
10.1007/s11760-022-02332-x
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep convolutional neural networks perform well in the field of computer vision, but exhibit undesirable behaviors such as memorization and sensitivity to adversarial examples. Therefore, proper regularization strategies are needed to alleviate these problems. Currently, regularization strategies with mixed sample data augmentation perform very well, and these algorithms allow the network to generalize better, improve the baseline performance of the model. However, interpolation-based mixed sample data augmentation distorts the data distribution, while masking-based mixed sample data augmentation results in excessive information loss for overly regular shapes of masks. Although mixed sample data augmentation is proven to be an effective method to improve the baseline performance, generalization ability and robustness of deep convolutional models, there is still room for improvement in terms of maintaining the of image local consistency and image data distribution. In this paper, we propose a new mixed sample data augmentation-LMix, which uses random masking to increase the number of masks in the image to maintain the data distribution, and high-frequency filtering to sharpen the image to highlight recognition regions. We applied the method to train CIFAR-10, CIFAR-100, SVHN, and Tiny-ImageNet datasets under the PreAct-ResNet18 model to evaluate the method, and obtained the latest results of 96.32, 79.85, 97.01, and 64.16%, respectively, which are 1.70, 4.73, and 8.06% higher than the optimal baseline accuracy. The LMix algorithm improves the generalization ability of the state-of-the-art neural network architecture and enhances the robustness to adversarial samples.
引用
收藏
页码:1245 / 1253
页数:9
相关论文
共 50 条
  • [1] LMix: regularization strategy for convolutional neural networks
    Linyu Yan
    Kunpeng Zheng
    Jinyao Xia
    Ke Li
    Hefei Ling
    Signal, Image and Video Processing, 2023, 17 : 1245 - 1253
  • [2] Weight Rotation as a Regularization Strategy in Convolutional Neural Networks
    Castro, Eduardo
    Pereira, Jose Costa
    Cardoso, Jaime S.
    2019 41ST ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY (EMBC), 2019, : 2106 - 2110
  • [3] Convolutional Neural Networks With Dynamic Regularization
    Wang, Yi
    Bian, Zhen-Peng
    Hou, Junhui
    Chau, Lap-Pui
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (05) : 2299 - 2304
  • [4] Multiscale Conditional Regularization for Convolutional Neural Networks
    Lu, Yao
    Lu, Guangming
    Li, Jinxing
    Xu, Yuanrong
    Zhang, Zheng
    Zhang, David
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (01) : 444 - 458
  • [5] REGULARIZATION OF CONVOLUTIONAL NEURAL NETWORKS USING SHUFFLENODE
    Chen, Yihao
    Wang, Hanli
    Long, Yu
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 355 - 360
  • [6] On the regularization of convolutional kernel tensors in neural networks
    Guo, Pei-Chang
    Ye, Qiang
    LINEAR & MULTILINEAR ALGEBRA, 2022, 70 (12): : 2318 - 2330
  • [7] TARGETDROP: A TARGETED REGULARIZATION METHOD FOR CONVOLUTIONAL NEURAL NETWORKS
    Zhu, Hui
    Zhao, Xiaofang
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 3283 - 3287
  • [8] A baseline regularization scheme for transfer learning with convolutional neural networks
    Li, Xuhong
    Grandvalet, Yves
    Davoine, Franck
    PATTERN RECOGNITION, 2020, 98
  • [9] The Effects of Regularization on Learning Facial Expressions with Convolutional Neural Networks
    Hinz, Tobias
    Barros, Pablo
    Wermter, Stefan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2016, PT II, 2016, 9887 : 80 - 87
  • [10] Avoiding Overfitting: A Survey on Regularization Methods for Convolutional Neural Networks
    Goncalves Dos Santos, Claudio Filipi
    Papa, Joao Paulo
    ACM COMPUTING SURVEYS, 2022, 54 (10S)