NearUni: Near-Unitary Training for Efficient Optical Neural Networks

被引:1
|
作者
Eldebiky, Amro [1 ]
Li, Bing [1 ]
Zhang, Grace Li [2 ]
机构
[1] Tech Univ Munich, Munich, Germany
[2] Tech Univ Darmstadt, Darmstadt, Germany
关键词
COMPACT;
D O I
10.1109/ICCAD57390.2023.10323877
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optical neural networks with Mach-Zender interferometers (MZIs) have demonstrated advantages over their electronic counterparts in computing efficiency and power consumption. However, implementing the computation with a weight matrix in DNNs using this technique requires the decomposition of the weight matrix into two unitary matrices, because an optical network can only realize a single unitary matrix due to its structural property. Accordingly, a direct implementation of DNNs onto optical networks suffer from a low area efficiency. To address this challenge, in this paper, a near-unitary training framework is proposed. In this framework, a weight matrix in DNNs is first partitioned into square submatrices to reduce the number of MZIs in the optical networks. Afterwards, training is adjusted to make the partitioned submatrices as close to unitary as possible. Such a matrix is then represented further by the sum of a unitary matrix and a sparse matrix. The latter implements the difference between the unitary matrix and the near-unitary matrix after training. In this way, only one optical network is needed to implement this unitary matrix and the low computation load in the sparse matrix can be implemented with area-efficient microring resonators (MRRs). Experimental results show that the area footprint can be reduced by 81.81%, 85.51%, 48.6% for ResNet34, VGG16, and fully connected neural networks, respectively, while the inference accuracy is still maintained on CIFAR100 and MNIST datasets.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [32] Efficient Training of Low-Curvature Neural Networks
    Srinivas, Suraj
    Matoba, Kyle
    Lakkaraju, Himabindu
    Fleuret, Francois
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [33] Efficient Communications in Training Large Scale Neural Networks
    Zhao, Yiyang
    Wang, Linnan
    Wu, Wei
    Bosilca, George
    Vuduc, Richard
    Ye, Jinmian
    Tang, Wenqi
    Xu, Zenglin
    PROCEEDINGS OF THE THEMATIC WORKSHOPS OF ACM MULTIMEDIA 2017 (THEMATIC WORKSHOPS'17), 2017, : 110 - 116
  • [34] Accurate, efficient and scalable training of Graph Neural Networks
    Zeng, Hanqing
    Zhou, Hongkuan
    Srivastava, Ajitesh
    Kannan, Rajgopal
    Prasanna, Viktor
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2021, 147 : 166 - 183
  • [35] Efficient EM training algorithm for probability neural networks
    Xiong, Hanchun
    He, Qianhua
    Li, Haizhou
    Huanan Ligong Daxue Xuebao/Journal of South China University of Technology (Natural Science), 1998, 26 (07): : 25 - 32
  • [36] Efficient training of large neural networks for language modeling
    Schwenk, H
    2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 3059 - 3064
  • [37] Efficient Training of Graph Neural Networks on Large Graphs
    Shen, Yanyan
    Chen, Lei
    Fang, Jingzhi
    Zhang, Xin
    Gao, Shihong
    Yin, Hongbo
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2024, 17 (12): : 4237 - 4240
  • [38] Efficient training of RBF neural networks for pattern recognition
    Lampariello, F
    Sciandrone, M
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2001, 12 (05): : 1235 - 1242
  • [39] Fully forward mode training for optical neural networks
    Xue, Zhiwei
    Zhou, Tiankuang
    Xu, Zhihao
    Yu, Shaoliang
    Dai, Qionghai
    Fang, Lu
    NATURE, 2024, 632 (8024) : 280 - 286
  • [40] Unitary Evolution Recurrent Neural Networks
    Arjovsky, Martin
    Shah, Amar
    Bengio, Yoshua
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48