NearUni: Near-Unitary Training for Efficient Optical Neural Networks

被引:1
|
作者
Eldebiky, Amro [1 ]
Li, Bing [1 ]
Zhang, Grace Li [2 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] Tech Univ Darmstadt, Darmstadt, Germany
关键词
COMPACT;
D O I
10.1109/ICCAD57390.2023.10323877
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optical neural networks with Mach-Zender interferometers (MZIs) have demonstrated advantages over their electronic counterparts in computing efficiency and power consumption. However, implementing the computation with a weight matrix in DNNs using this technique requires the decomposition of the weight matrix into two unitary matrices, because an optical network can only realize a single unitary matrix due to its structural property. Accordingly, a direct implementation of DNNs onto optical networks suffer from a low area efficiency. To address this challenge, in this paper, a near-unitary training framework is proposed. In this framework, a weight matrix in DNNs is first partitioned into square submatrices to reduce the number of MZIs in the optical networks. Afterwards, training is adjusted to make the partitioned submatrices as close to unitary as possible. Such a matrix is then represented further by the sum of a unitary matrix and a sparse matrix. The latter implements the difference between the unitary matrix and the near-unitary matrix after training. In this way, only one optical network is needed to implement this unitary matrix and the low computation load in the sparse matrix can be implemented with area-efficient microring resonators (MRRs). Experimental results show that the area footprint can be reduced by 81.81%, 85.51%, 48.6% for ResNet34, VGG16, and fully connected neural networks, respectively, while the inference accuracy is still maintained on CIFAR100 and MNIST datasets.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] Deep Unitary Convolutional Neural Networks
    Chang, Hao-Yuan
    Wang, Kang L.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 170 - 181
  • [42] Efficient Trainability of Linear Optical Modules in Quantum Optical Neural Networks
    Volkoff, Tyler J.
    JOURNAL OF RUSSIAN LASER RESEARCH, 2021, 42 (03) : 250 - 260
  • [43] Efficient Trainability of Linear Optical Modules in Quantum Optical Neural Networks
    Tyler J. Volkoff
    Journal of Russian Laser Research, 2021, 42 : 250 - 260
  • [44] On unitary reconstruction of linear optical networks
    Tillmann, Max
    Schmidt, Christian
    Walther, Philip
    JOURNAL OF OPTICS, 2016, 18 (11)
  • [45] Design and Training of Binarized Neural Networks for Highly Efficient Accelerators
    Li J.
    Xu H.
    Wang Y.
    Xiao H.
    Wang Y.
    Han Y.
    Li X.
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2023, 35 (06): : 961 - 969
  • [46] Memory-Efficient Training of Binarized Neural Networks on the Edge
    Yayla, Mikail
    Chen, Jian-Jia
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 661 - 666
  • [47] Efficient Training of Very Deep Neural Networks for Supervised Hashing
    Zhang, Ziming
    Chen, Yuting
    Saligrama, Venkatesh
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1487 - 1495
  • [48] GradInit: Learning to Initialize Neural Networks for Stable and Efficient Training
    Zhu, Chen
    Ni, Renkun
    Xu, Zheng
    Kong, Kezhi
    Huang, W. Ronny
    Goldstein, Tom
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [49] An Efficient Approach to Escalate the Speed of Training Convolution Neural Networks
    P Pabitha
    Anusha Jayasimhan
    China Communications, 2024, 21 (02) : 258 - 269
  • [50] An Efficient Approach to Escalate the Speed of Training Convolution Neural Networks
    Pabitha, P.
    Jayasimhan, Anusha
    CHINA COMMUNICATIONS, 2024, 21 (02) : 258 - 269