NearUni: Near-Unitary Training for Efficient Optical Neural Networks

被引:1
|
作者
Eldebiky, Amro [1 ]
Li, Bing [1 ]
Zhang, Grace Li [2 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] Tech Univ Darmstadt, Darmstadt, Germany
关键词
COMPACT;
D O I
10.1109/ICCAD57390.2023.10323877
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optical neural networks with Mach-Zender interferometers (MZIs) have demonstrated advantages over their electronic counterparts in computing efficiency and power consumption. However, implementing the computation with a weight matrix in DNNs using this technique requires the decomposition of the weight matrix into two unitary matrices, because an optical network can only realize a single unitary matrix due to its structural property. Accordingly, a direct implementation of DNNs onto optical networks suffer from a low area efficiency. To address this challenge, in this paper, a near-unitary training framework is proposed. In this framework, a weight matrix in DNNs is first partitioned into square submatrices to reduce the number of MZIs in the optical networks. Afterwards, training is adjusted to make the partitioned submatrices as close to unitary as possible. Such a matrix is then represented further by the sum of a unitary matrix and a sparse matrix. The latter implements the difference between the unitary matrix and the near-unitary matrix after training. In this way, only one optical network is needed to implement this unitary matrix and the low computation load in the sparse matrix can be implemented with area-efficient microring resonators (MRRs). Experimental results show that the area footprint can be reduced by 81.81%, 85.51%, 48.6% for ResNet34, VGG16, and fully connected neural networks, respectively, while the inference accuracy is still maintained on CIFAR100 and MNIST datasets.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] In situ optical backpropagation training of diffractive optical neural networks
    TIANKUANG ZHOU
    LU FANG
    TAO YAN
    JIAMIN WU
    YIPENG LI
    JINGTAO FAN
    HUAQIANG WU
    XING LIN
    QIONGHAI DAI
    Photonics Research, 2020, 8 (06) : 940 - 953
  • [22] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    Advances in Neural Information Processing Systems, 2022, 35
  • [23] Efficient and effective training of sparse recurrent neural networks
    Shiwei Liu
    Iftitahu Ni’mah
    Vlado Menkovski
    Decebal Constantin Mocanu
    Mykola Pechenizkiy
    Neural Computing and Applications, 2021, 33 : 9625 - 9636
  • [24] Efficient Incremental Training for Deep Convolutional Neural Networks
    Tao, Yudong
    Tu, Yuexuan
    Shyu, Mei-Ling
    2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, : 286 - 291
  • [25] An Efficient Optimization Technique for Training Deep Neural Networks
    Mehmood, Faisal
    Ahmad, Shabir
    Whangbo, Taeg Keun
    MATHEMATICS, 2023, 11 (06)
  • [26] An efficient global algorithm for supervised training of neural networks
    Shukla, KK
    Raghunath
    COMPUTERS & ELECTRICAL ENGINEERING, 1999, 25 (03) : 193 - 216
  • [27] EXODUS: Stable and efficient training of spiking neural networks
    Bauer, Felix C.
    Lenz, Gregor
    Haghighatshoar, Saeid
    Sheik, Sadique
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [28] Efficient Constructive Techniques for Training Switching Neural Networks
    Ferrari, Enrico
    Muselli, Marco
    CONSTRUCTIVE NEURAL NETWORKS, 2009, 258 : 25 - 48
  • [29] Efficient Training of Artificial Neural Networks for Autonomous Navigation
    Pomerleau, Dean A.
    NEURAL COMPUTATION, 1991, 3 (01) : 88 - 97
  • [30] Efficient and effective training of sparse recurrent neural networks
    Liu, Shiwei
    Ni'mah, Iftitahu
    Menkovski, Vlado
    Mocanu, Decebal Constantin
    Pechenizkiy, Mykola
    NEURAL COMPUTING & APPLICATIONS, 2021, 33 (15): : 9625 - 9636