Efficient on-chip training of large-scale optical neural network through block adjoint training algorithm

被引:0
|
作者
Yang, Zhiwei [1 ,2 ]
Zhang, Tian [1 ,2 ]
Dai, Jian [1 ,2 ]
Xu, Kun [1 ,2 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Informat Photon & Opt Commun, Beijing 100876, Peoples R China
[2] Beijing Univ Posts & Telecommun, Sch Elect Engn, Beijing 100876, Peoples R China
来源
OPTICS EXPRESS | 2024年 / 32卷 / 26期
基金
中国国家自然科学基金;
关键词
DESIGN;
D O I
10.1364/OE.537813
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
MZI-based block optical neural networks (BONNs), which utilize block matrix multiplication to achieve large-scale network models, have garnered significant attention but still lack efficient training algorithms. In this article, by calculating the original field and adjoint field for the block matrices in BONNs and directly updating the phase values of all phase shifters within the optical mesh, we propose an on-chip block adjoint training (BAT) algorithm for large-scale BONNs. To demonstrate the effectiveness of our proposed algorithm, the trained BONNs are applied in image classification tasks for MNIST and SVHN datasets. The calculated results demonstrate that the performance of the BAT algorithm (95.915% for the MNIST dataset and 82.64% for the SVHN dataset) is competitive with the traditional gradient algorithm based on artificial neural networks (96.238% and 84.182%), but the BONNs can infer 1.5 times and 1.3 times faster than artificial neural networks, respectively. By studying the influence of the block size and the inputted position of the padded zero signals, we demonstrate that the BAT algorithm based on the BONNs with 12 block sizes can achieve higher performance by adding the padded zero signals to the same side beside the normal inputted signals. Additionally, we demonstrate that substituting the complete weight matrices with unitary matrices to construct BONNs is an efficient way to reduce both the system area and the required trainable parameters. Finally, we demonstrate the relatively good robustness of the BAT algorithm and the imprecision alleviation method by using on-chip retraining. Notably, our proposed BAT algorithm shows excellent potential for more complex tasks and network models.
引用
收藏
页码:46633 / 46648
页数:16
相关论文
共 50 条
  • [41] Efficient training and design of photonic neural network through neuroevolution
    Zhang, Tian
    Wang, Jia
    Dan, Yihang
    Lanqiu, Yuxiang
    Dai, Jian
    Han, Xu
    Sun, Xiaojuan
    Xu, Kun
    OPTICS EXPRESS, 2019, 27 (26): : 37150 - 37163
  • [42] SNNOT: Spiking Neural Network With On-Chip Training for MIMO-OFDM Symbol Detection
    Zheng, Honghao
    Xu, Jiarui
    Liu, Lingjia
    Yi, Yang
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2024, 8 (04): : 1809 - 1823
  • [43] On-Chip Communication Network for Efficient Training of Deep Convolutional Networks on Heterogeneous Manycore Systems
    Choi, Wonje
    Duraisamy, Karthi
    Kim, Ryan Gary
    Doppa, Janardhan Rao
    Pande, Partha Pratim
    Marculescu, Diana
    Marculescu, Radu
    IEEE TRANSACTIONS ON COMPUTERS, 2018, 67 (05) : 672 - 686
  • [44] Unidirectional and hierarchical on-chip interconnected architecture for large-scale hardware spiking neural networks
    Liu, Junxiu
    Jiang, Dong
    Fu, Qiang
    Luo, Yuling
    Deng, Yaohua
    Qin, Sheng
    Zhang, Shunsheng
    NEUROCOMPUTING, 2024, 609
  • [45] Chimera: Efficiently Training Large-Scale Neural Networks with Bidirectional Pipelines
    Li, Shigang
    Hoefler, Torsten
    SC21: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2021,
  • [46] TAG: A Neural Network Model for Large-Scale Optical Implementation
    Lee, Hyuek-Jae
    Lee, Soo-Young
    Shin, Sang-Yung
    Koh, Bo-Yun
    NEURAL COMPUTATION, 1991, 3 (01) : 135 - 143
  • [47] Rethinking residual connection in training large-scale spiking neural networks
    Li, Yudong
    Lei, Yunlin
    Yang, Xu
    NEUROCOMPUTING, 2025, 616
  • [48] MLPs: Efficient Training of MiniGo on Large-scale Heterogeneous Computing System
    Qiao, Peng
    He, Zhouyu
    Li, Rongchun
    Jiang, Jingfei
    Dou, Yong
    Li, Dongsheng
    2022 IEEE 28TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, ICPADS, 2022, : 475 - 482
  • [49] Resource-efficient Federated Learning for Large-scale Model Training
    Song, Zilin
    Li, Zhengze
    Yuan, Tingting
    Fu, Xiaoming
    PROCEEDINGS OF THE WORKSHOP ON MOBILITY IN THE EVOLVING INTERNET ARCHITECTURE TO BE HELD IN CONJUNCTION WITH MOBICOM 2024, MOBIARCH 2024, 2024, : 43 - 48
  • [50] MixPipe: Efficient Bidirectional Pipeline Parallelism for Training Large-Scale Models
    Zhang, Weigang
    Zhou, Biyu
    Tang, Xuehai
    Wang, Zhaoxing
    Hu, Songlin
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,