Post-training approach for mitigating overfitting in quantum convolutional neural networks

被引:0
|
作者
Shinde, Aakash Ravindra [1 ]
Jain, Charu [1 ]
Kalev, Amir [2 ,3 ,4 ]
机构
[1] Univ Southern Calif, Grad Sch, Viterbi Sch Engn, Los Angeles, CA 90089 USA
[2] Univ Southern Calif, Informat Sci Inst, Arlington, VA 22203 USA
[3] Univ Southern Calif, Dept Phys & Astron, Los Angeles, CA 90089 USA
[4] Univ Southern Calif, Ctr Quantum Informat Sci & Technol, Los Angeles, CA 90089 USA
关键词
Convolution - Convolutional neural networks - Quantum entanglement;
D O I
10.1103/PhysRevA.110.042409
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Quantum convolutional neural network (QCNN), an early application for quantum computers in the noisy intermediate-scale quantum era, has been consistently proven successful as a machine learning (ML) algorithm for several tasks with significant accuracy. Derived from its classical counterpart, QCNN is prone to overfitting. Overfitting is a typical shortcoming of ML models that are trained too closely to the availed training dataset and perform relatively poorly on unseen datasets for a similar problem. In this work we study post-training approaches for mitigating overfitting in QCNNs. We find that a straightforward adaptation of a classical post-training method, known as neuron dropout, to the quantum setting leads to a significant and undesirable consequence: a substantial decrease in success probability of the QCNN. We argue that this effect exposes the crucial role of entanglement in QCNNs and the vulnerability of QCNNs to entanglement loss. Hence, we propose a parameter adaptation method as an alternative method. Our method is computationally efficient and is found to successfully handle overfitting in the test cases.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Post-training on RBF neural networks
    Shabaninia, Faridoon
    Roopaei, Mehdi
    Fatemi, Mehdi
    NONLINEAR ANALYSIS-HYBRID SYSTEMS, 2007, 1 (04) : 491 - 500
  • [2] Normalized Post-training Quantization for Photonic Neural Networks
    Kirtas, M.
    Passalis, N.
    Oikonomou, A.
    Mourgias-Alexandris, G.
    Moralis-Pegios, M.
    Pleros, N.
    Tefas, A.
    2022 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2022, : 657 - 663
  • [3] Post-training Quantization for Neural Networks with Provable Guarantees*
    Zhang, Jinjie
    Zhou, Yixuan
    Saab, Rayan
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (02): : 373 - 399
  • [4] Quantune: Post-training quantization of convolutional neural networks using extreme gradient boosting for fast deployment
    Lee, Jemin
    Yu, Misun
    Kwon, Yongin
    Kim, Taeho
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2022, 132 : 124 - 135
  • [5] Optimizing Convolutional Neural Network Performance by Mitigating Underfitting and Overfitting
    Li, Qipei
    Yan, Ming
    Xu, Jie
    2021 IEEE/ACIS 20TH INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION SCIENCE (ICIS 2021-SUMMER), 2021, : 126 - 131
  • [6] PTMQ: Post-training Multi-Bit Quantization of Neural Networks
    Xu, Ke
    Li, Zhongcheng
    Wang, Shanshan
    Zhang, Xingyi
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 16193 - 16201
  • [7] Post-Training Quantization for Energy Efficient Realization of Deep Neural Networks
    Latotzke, Cecilia
    Balim, Batuhan
    Gemmeke, Tobias
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 1559 - 1566
  • [8] Lost-minimum post-training parameter quantization method for convolutional neural network
    Zhang F.
    Huang Y.
    Fang Z.
    Guo W.
    Tongxin Xuebao/Journal on Communications, 2022, 43 (04): : 114 - 122
  • [9] Effective Post-Training Quantization Of Neural Networks For Inference on Low Power Neural Accelerator
    Demidovskij, Alexander
    Smirnov, Eugene
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [10] Cropout: A General Mechanism for Reducing Overfitting on Convolutional Neural Networks
    Hou, Wenbo
    Wang, Wenhai
    Liu, Ruo-Ze
    Lu, Tong
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,