Post-training approach for mitigating overfitting in quantum convolutional neural networks

被引:0
|
作者
Shinde, Aakash Ravindra [1 ]
Jain, Charu [1 ]
Kalev, Amir [2 ,3 ,4 ]
机构
[1] Univ Southern Calif, Grad Sch, Viterbi Sch Engn, Los Angeles, CA 90089 USA
[2] Univ Southern Calif, Informat Sci Inst, Arlington, VA 22203 USA
[3] Univ Southern Calif, Dept Phys & Astron, Los Angeles, CA 90089 USA
[4] Univ Southern Calif, Ctr Quantum Informat Sci & Technol, Los Angeles, CA 90089 USA
关键词
Convolution - Convolutional neural networks - Quantum entanglement;
D O I
10.1103/PhysRevA.110.042409
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Quantum convolutional neural network (QCNN), an early application for quantum computers in the noisy intermediate-scale quantum era, has been consistently proven successful as a machine learning (ML) algorithm for several tasks with significant accuracy. Derived from its classical counterpart, QCNN is prone to overfitting. Overfitting is a typical shortcoming of ML models that are trained too closely to the availed training dataset and perform relatively poorly on unseen datasets for a similar problem. In this work we study post-training approaches for mitigating overfitting in QCNNs. We find that a straightforward adaptation of a classical post-training method, known as neuron dropout, to the quantum setting leads to a significant and undesirable consequence: a substantial decrease in success probability of the QCNN. We argue that this effect exposes the crucial role of entanglement in QCNNs and the vulnerability of QCNNs to entanglement loss. Hence, we propose a parameter adaptation method as an alternative method. Our method is computationally efficient and is found to successfully handle overfitting in the test cases.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] JOINT TRAINING OF CONVOLUTIONAL AND NON-CONVOLUTIONAL NEURAL NETWORKS
    Soltau, Hagen
    Saon, George
    Sainath, Tara N.
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [42] Scaling up the training of Convolutional Neural Networks
    Snir, Marc
    2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 925 - 925
  • [43] Towards dropout training for convolutional neural networks
    Wu, Haibing
    Gu, Xiaodong
    NEURAL NETWORKS, 2015, 71 : 1 - 10
  • [44] CONVOLUTIONAL NEURAL NETWORKS TRAINING FOR AUTONOMOUS ROBOTICS
    Lozhkin, Alexander
    Maiorov, Konstantin
    Bozek, Pavol
    MANAGEMENT SYSTEMS IN PRODUCTION ENGINEERING, 2021, 29 (01) : 75 - 79
  • [45] Quantum optimization for training quantum neural networks
    Liao, Yidong
    Hsieh, Min-Hsiu
    Ferrie, Chris
    QUANTUM MACHINE INTELLIGENCE, 2024, 6 (01)
  • [46] A Tutorial on Quantum Convolutional Neural Networks (QCNN)
    Oh, Seunghyeok
    Choi, Jaeho
    Kim, Joongheon
    11TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE: DATA, NETWORK, AND AI IN THE AGE OF UNTACT (ICTC 2020), 2020, : 236 - 239
  • [47] Channel attention for quantum convolutional neural networks
    Budiutama, Gekko
    Daimon, Shunsuke
    Nishi, Hirofumi
    Kaneko, Ryui
    Ohtsuki, Tomi
    Matsushita, Yu-ichiro
    PHYSICAL REVIEW A, 2024, 110 (01)
  • [48] Stereoscopic scalable quantum convolutional neural networks
    Baek, Hankyul
    Yun, Won Joon
    Park, Soohyun
    Kim, Joongheon
    NEURAL NETWORKS, 2023, 165 : 860 - 867
  • [49] Quantum Similarity Testing with Convolutional Neural Networks
    Wu, Ya-Dong
    Zhu, Yan
    Bai, Ge
    Wang, Yuexuan
    Chiribella, Giulio
    PHYSICAL REVIEW LETTERS, 2023, 130 (21)
  • [50] Hybrid Post-Training Quantization for Super-Resolution Neural Network Compression
    Xu, Naijie
    Chen, Xiaohui
    Cao, Youlong
    Zhang, Wenyi
    IEEE SIGNAL PROCESSING LETTERS, 2023, 30 : 379 - 383