Post-training approach for mitigating overfitting in quantum convolutional neural networks

被引:0
|
作者
Shinde, Aakash Ravindra [1 ]
Jain, Charu [1 ]
Kalev, Amir [2 ,3 ,4 ]
机构
[1] Univ Southern Calif, Grad Sch, Viterbi Sch Engn, Los Angeles, CA 90089 USA
[2] Univ Southern Calif, Informat Sci Inst, Arlington, VA 22203 USA
[3] Univ Southern Calif, Dept Phys & Astron, Los Angeles, CA 90089 USA
[4] Univ Southern Calif, Ctr Quantum Informat Sci & Technol, Los Angeles, CA 90089 USA
关键词
Convolution - Convolutional neural networks - Quantum entanglement;
D O I
10.1103/PhysRevA.110.042409
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Quantum convolutional neural network (QCNN), an early application for quantum computers in the noisy intermediate-scale quantum era, has been consistently proven successful as a machine learning (ML) algorithm for several tasks with significant accuracy. Derived from its classical counterpart, QCNN is prone to overfitting. Overfitting is a typical shortcoming of ML models that are trained too closely to the availed training dataset and perform relatively poorly on unseen datasets for a similar problem. In this work we study post-training approaches for mitigating overfitting in QCNNs. We find that a straightforward adaptation of a classical post-training method, known as neuron dropout, to the quantum setting leads to a significant and undesirable consequence: a substantial decrease in success probability of the QCNN. We argue that this effect exposes the crucial role of entanglement in QCNNs and the vulnerability of QCNNs to entanglement loss. Hence, we propose a parameter adaptation method as an alternative method. Our method is computationally efficient and is found to successfully handle overfitting in the test cases.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Latent Training for Convolutional Neural Networks
    Huang, Zi
    Liu, Qi
    Chen, Zhiyuan
    Zhao, Yuming
    PROCEEDINGS OF 2015 INTERNATIONAL CONFERENCE ON ESTIMATION, DETECTION AND INFORMATION FUSION ICEDIF 2015, 2015, : 55 - 60
  • [32] Post-training Iterative Hierarchical Data Augmentation for Deep Networks
    Khan, Adil
    Fraz, Khadija
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [33] Quantum Convolution for Convolutional Neural Networks
    Bourahla, Mustapha
    QUANTUM COMPUTING: APPLICATIONS AND CHALLENGES, QSAC 2023, 2024, 2 : 179 - 193
  • [34] Quantum Dilated Convolutional Neural Networks
    Chen, Yixiong
    IEEE ACCESS, 2022, 10 : 20240 - 20246
  • [35] Branching quantum convolutional neural networks
    MacCormack, Ian
    Delaney, Conor
    Galda, Alexey
    Aggarwal, Nidhi
    Narang, Prineha
    PHYSICAL REVIEW RESEARCH, 2022, 4 (01):
  • [36] Quantum Graph Convolutional Neural Networks
    Zheng, Jin
    Gao, Qing
    Lu, Yanxuan
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 6335 - 6340
  • [37] AE-Qdrop: Towards Accurate and Efficient Low-Bit Post-Training Quantization for A Convolutional Neural Network
    Li, Jixing
    Chen, Gang
    Jin, Min
    Mao, Wenyu
    Lu, Huaxiang
    ELECTRONICS, 2024, 13 (03)
  • [38] Handling Overfitting and Imbalance Data in Modelling Convolutional Neural Networks for Astronomical Transient Discovery
    Boongoen, Tossapon
    Iam-On, Natthakan
    CONTRIBUTIONS PRESENTED AT THE INTERNATIONAL CONFERENCE ON COMPUTING, COMMUNICATION, CYBERSECURITY AND AI, C3AI 2024, 2024, 884 : 691 - 698
  • [39] Improving the Post-Training Neural Network Quantization by Prepositive Feature Quantization
    Chu, Tianshu
    Yang, Zuopeng
    Huang, Xiaolin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (04) : 3056 - 3060
  • [40] Hybrid Quantum Convolutional Neural Networks in TensorFlow Quantum
    Khurelsukh, Bishrelt
    2022 33RD IRISH SIGNALS AND SYSTEMS CONFERENCE (ISSC), 2022,