Efficient on-chip training of large-scale optical neural network through block adjoint training algorithm

被引:0
|
作者
Yang, Zhiwei [1 ,2 ]
Zhang, Tian [1 ,2 ]
Dai, Jian [1 ,2 ]
Xu, Kun [1 ,2 ]
机构
[1] Beijing Univ Posts & Telecommun, State Key Lab Informat Photon & Opt Commun, Beijing 100876, Peoples R China
[2] Beijing Univ Posts & Telecommun, Sch Elect Engn, Beijing 100876, Peoples R China
来源
OPTICS EXPRESS | 2024年 / 32卷 / 26期
基金
中国国家自然科学基金;
关键词
DESIGN;
D O I
10.1364/OE.537813
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
MZI-based block optical neural networks (BONNs), which utilize block matrix multiplication to achieve large-scale network models, have garnered significant attention but still lack efficient training algorithms. In this article, by calculating the original field and adjoint field for the block matrices in BONNs and directly updating the phase values of all phase shifters within the optical mesh, we propose an on-chip block adjoint training (BAT) algorithm for large-scale BONNs. To demonstrate the effectiveness of our proposed algorithm, the trained BONNs are applied in image classification tasks for MNIST and SVHN datasets. The calculated results demonstrate that the performance of the BAT algorithm (95.915% for the MNIST dataset and 82.64% for the SVHN dataset) is competitive with the traditional gradient algorithm based on artificial neural networks (96.238% and 84.182%), but the BONNs can infer 1.5 times and 1.3 times faster than artificial neural networks, respectively. By studying the influence of the block size and the inputted position of the padded zero signals, we demonstrate that the BAT algorithm based on the BONNs with 12 block sizes can achieve higher performance by adding the padded zero signals to the same side beside the normal inputted signals. Additionally, we demonstrate that substituting the complete weight matrices with unitary matrices to construct BONNs is an efficient way to reduce both the system area and the required trainable parameters. Finally, we demonstrate the relatively good robustness of the BAT algorithm and the imprecision alleviation method by using on-chip retraining. Notably, our proposed BAT algorithm shows excellent potential for more complex tasks and network models.
引用
收藏
页码:46633 / 46648
页数:16
相关论文
共 50 条
  • [1] Efficient On-Chip Training of Optical Neural Networks Using Genetic Algorithm
    Zhang, Hui
    Thompson, Jayne
    Gu, Mile
    Jiang, Xu Dong
    Cai, Hong
    Liu, Patricia Yang
    Shi, Yuzhi
    Zhang, Yi
    Karim, Muhammad Faeyz
    Lo, Guo Qiang
    Luo, Xianshu
    Dong, Bin
    Kwek, Leong Chuan
    Liu, Ai Qun
    ACS PHOTONICS, 2021, 8 (06) : 1662 - 1672
  • [2] Local cluster neural network on-chip training
    Zhang, Liang
    Sitte, Joaquin
    2006 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORK PROCEEDINGS, VOLS 1-10, 2006, : 29 - +
  • [3] POSTER: ParGNN: Efficient Training for Large-Scale Graph Neural Network on GPU Clusters
    Li, Shunde
    Gu, Junyu
    Wang, Jue
    Yao, Tiechui
    Liang, Zhiqiang
    Shi, Yumeng
    Li, Shigang
    Xi, Weiting
    Li, Shushen
    Zhou, Chunbao
    Wang, Yangang
    Chi, Xuebin
    PROCEEDINGS OF THE 29TH ACM SIGPLAN ANNUAL SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, PPOPP 2024, 2024, : 469 - 471
  • [4] Recursive Binary Neural Network Training Model for Efficient Usage of On-Chip Memory
    Guan, Tianchan
    Liu, Peiye
    Zeng, Xiaoyang
    Kim, Martha
    Seok, Mingoo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2019, 66 (07) : 2593 - 2605
  • [5] ByteGNN: Efficient Graph Neural Network Training at Large Scale
    Zheng, Chenguang
    Chen, Hongzhi
    Cheng, Yuxuan
    Song, Zhezheng
    Wu, Yifan
    Li, Changji
    Cheng, James
    Yang, Hao
    Zhang, Shuai
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2022, 15 (06): : 1228 - 1242
  • [6] Parallel Large-Scale Neural Network Training For Online Advertising
    Qi, Quanchang
    Lu, Guangming
    Zhang, Jun
    Yang, Lichun
    Liu, Haishan
    2018 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2018, : 343 - 350
  • [7] An Optical Computing Chip for Executing Complex-valued Neural Network and Its On-chip Training
    Zhang, Hui
    Liu, Ai Qun
    2022 PHOTONICS & ELECTROMAGNETICS RESEARCH SYMPOSIUM (PIERS 2022), 2022, : 189 - 196
  • [8] NeuronLink: An Efficient Chip-to-Chip Interconnect for Large-Scale Neural Network Accelerators
    Xiao, Shanlin
    Guo, Yuhao
    Liao, Wenkang
    Deng, Huipeng
    Luo, Yi
    Zheng, Huanliang
    Wang, Jian
    Li, Cheng
    Li, Gezi
    Yu, Zhiyi
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (09) : 1966 - 1978
  • [9] NeuronLink: An Efficient Chip-to-Chip Interconnect for Large-Scale Neural Network Accelerators
    Xiao, Shanlin
    Guo, Yuhao
    Liao, Wenkang
    Deng, Huipeng
    Luo, Yi
    Zheng, Huanliang
    Wang, Jian
    Li, Cheng
    Li, Gezi
    Yu, Zhiyi
    Xiao, Shanlin (xiaoshlin@mail.sysu.edu.cn); Yu, Zhiyi (yuzhiyi@mail.sysu.edu.cn), 1966, Institute of Electrical and Electronics Engineers Inc. (28): : 1966 - 1978
  • [10] FEDOBD: Opportunistic Block Dropout for Efficiently Training Large-scale Neural Networks through Federated Learning
    Chen, Yuanyuan
    Chen, Zichen
    Yu, Pengcheng
    Yu, Han
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 3541 - 3549