Fully forward mode training for optical neural networks

被引:6
|
作者
Xue, Zhiwei [1 ,2 ,3 ,4 ]
Zhou, Tiankuang [1 ,2 ,3 ]
Xu, Zhihao [1 ,2 ,3 ,4 ]
Yu, Shaoliang [5 ]
Dai, Qionghai [2 ,3 ,6 ]
Fang, Lu [1 ,2 ,3 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
[2] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, Beijing, Peoples R China
[3] Tsinghua Univ, Inst Brain & Cognit Sci, Beijing, Peoples R China
[4] Tsinghua Univ, Shenzhen Int Grad Sch, Shenzhen, Peoples R China
[5] Zhejiang Lab, Res Ctr Intelligent Optoelect Comp, Hangzhou, Peoples R China
[6] Tsinghua Univ, Dept Automat, Beijing, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
INVERSE DESIGN; ARTIFICIAL-INTELLIGENCE; BACKPROPAGATION; TIME;
D O I
10.1038/s41586-024-07687-4
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Optical computing promises to improve the speed and energy efficiency of machine learning applications1-6. However, current approaches to efficiently train these models are limited by in silico emulation on digital computers. Here we develop a method called fully forward mode (FFM) learning, which implements the compute-intensive training process on the physical system. The majority of the machine learning operations are thus efficiently conducted in parallel on site, alleviating numerical modelling constraints. In free-space and integrated photonics, we experimentally demonstrate optical systems with state-of-the-art performances for a given network size. FFM learning shows training the deepest optical neural networks with millions of parameters achieves accuracy equivalent to the ideal model. It supports all-optical focusing through scattering media with a resolution of the diffraction limit; it can also image in parallel the objects hidden outside the direct line of sight at over a kilohertz frame rate and can conduct all-optical processing with light intensity as weak as subphoton per pixel (5.40 x 1018- operations-per-second-per-watt energy efficiency) at room temperature. Furthermore, we prove that FFM learning can automatically search non-Hermitian exceptional points without an analytical model. FFM learning not only facilitates orders-of-magnitude-faster learning processes, but can also advance applied and theoretical fields such as deep neural networks, ultrasensitive perception and topological photonics. We present fully forward mode learning, which conducts machine learning operations on site, leading to faster learning and promoting advancement in numerous fields.
引用
收藏
页码:280 / 286
页数:17
相关论文
共 50 条
  • [31] Sliding mode algorithm for training multilayer artificial neural networks
    Parma, GG
    Menezes, BR
    Braga, AP
    ELECTRONICS LETTERS, 1998, 34 (01) : 97 - 98
  • [32] Free Probability for predicting the performance of feed-forward fully connected neural networks
    Chhaibi, Reda
    Daouda, Tariq
    Kahn, Ezechiel
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [33] RTM Gravity Forward Modeling Using Improved Fully Connected Deep Neural Networks
    Zhang, Baoyu
    Yang, Meng
    Feng, Wei
    Jiang, Mi
    Yan, Xinyuan
    Zhong, Min
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2024, 62
  • [34] Free Probability for predicting the performance of feed-forward fully connected neural networks
    Chhaibi, Reda
    Daouda, Tariq
    Kahn, Ezéchiel
    arXiv, 2021,
  • [35] Free Probability for predicting the performance of feed-forward fully connected neural networks
    Chhaibi, Reda
    Daouda, Tariq
    Kahn, Ezéchiel
    Advances in Neural Information Processing Systems, 2022, 35
  • [36] On the BP Training Algorithm of Fuzzy Neural Networks (FNNs) via Its Equivalent Fully Connected Neural Networks (FFNNs)
    Wang, Jing
    Wang, Chi-Hsu
    Chen, C. L. Philip
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 1376 - 1381
  • [37] Differentiable homotopy methods for gradually reinforcing the training of fully connected neural networks
    Li, Peixuan
    Li, Yuanbo
    NEUROCOMPUTING, 2024, 605
  • [38] Distributed Learning of Fully Connected Neural Networks using Independent Subnet Training
    Yuan, Binhang
    Wolfe, Cameron R.
    Dun, Chen
    Tang, Yuxin
    Kyrillidis, Anastasios
    Jermaine, Chris
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2022, 15 (08): : 1581 - 1590
  • [39] Dynamic group optimisation algorithm for training feed-forward neural networks
    Tang, Rui
    Fong, Simon
    Deb, Suash
    Vasilakos, Athanasios V.
    Millham, Richard C.
    NEUROCOMPUTING, 2018, 314 : 1 - 19
  • [40] Unsupervised, smooth training of feed-forward neural networks for mismatch compensation
    Surendran, AC
    Lee, CH
    Rahim, M
    1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, PROCEEDINGS, 1997, : 482 - 489