Fully forward mode training for optical neural networks

被引:6
|
作者
Xue, Zhiwei [1 ,2 ,3 ,4 ]
Zhou, Tiankuang [1 ,2 ,3 ]
Xu, Zhihao [1 ,2 ,3 ,4 ]
Yu, Shaoliang [5 ]
Dai, Qionghai [2 ,3 ,6 ]
Fang, Lu [1 ,2 ,3 ]
机构
[1] Tsinghua Univ, Dept Elect Engn, Beijing, Peoples R China
[2] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, Beijing, Peoples R China
[3] Tsinghua Univ, Inst Brain & Cognit Sci, Beijing, Peoples R China
[4] Tsinghua Univ, Shenzhen Int Grad Sch, Shenzhen, Peoples R China
[5] Zhejiang Lab, Res Ctr Intelligent Optoelect Comp, Hangzhou, Peoples R China
[6] Tsinghua Univ, Dept Automat, Beijing, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
INVERSE DESIGN; ARTIFICIAL-INTELLIGENCE; BACKPROPAGATION; TIME;
D O I
10.1038/s41586-024-07687-4
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Optical computing promises to improve the speed and energy efficiency of machine learning applications1-6. However, current approaches to efficiently train these models are limited by in silico emulation on digital computers. Here we develop a method called fully forward mode (FFM) learning, which implements the compute-intensive training process on the physical system. The majority of the machine learning operations are thus efficiently conducted in parallel on site, alleviating numerical modelling constraints. In free-space and integrated photonics, we experimentally demonstrate optical systems with state-of-the-art performances for a given network size. FFM learning shows training the deepest optical neural networks with millions of parameters achieves accuracy equivalent to the ideal model. It supports all-optical focusing through scattering media with a resolution of the diffraction limit; it can also image in parallel the objects hidden outside the direct line of sight at over a kilohertz frame rate and can conduct all-optical processing with light intensity as weak as subphoton per pixel (5.40 x 1018- operations-per-second-per-watt energy efficiency) at room temperature. Furthermore, we prove that FFM learning can automatically search non-Hermitian exceptional points without an analytical model. FFM learning not only facilitates orders-of-magnitude-faster learning processes, but can also advance applied and theoretical fields such as deep neural networks, ultrasensitive perception and topological photonics. We present fully forward mode learning, which conducts machine learning operations on site, leading to faster learning and promoting advancement in numerous fields.
引用
收藏
页码:280 / 286
页数:17
相关论文
共 50 条
  • [1] Forward-forward training of an optical neural network
    Oguz, Ilker
    Ke, Junjie
    Weng, Qifei
    Yang, Feng
    Yildirim, Mustafa
    Dinc, Niyazi Ulas
    Hsieh, Jih-Liang
    Moser, Christophe
    Psaltis, Demetri
    OPTICS LETTERS, 2023, 48 (20) : 5249 - 5252
  • [2] TRAINING FULLY RECURRENT NEURAL NETWORKS WITH COMPLEX WEIGHTS
    KECHRIOTIS, G
    MANOLAKOS, ES
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-ANALOG AND DIGITAL SIGNAL PROCESSING, 1994, 41 (03): : 235 - 238
  • [3] A homotopy training algorithm for fully connected neural networks
    Chen, Qipin
    Hao, Wenrui
    PROCEEDINGS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2019, 475 (2231):
  • [4] Hybrid training of optical neural networks
    Spall, James
    Guo, Xianxin
    Lvovsky, A., I
    OPTICA, 2022, 9 (07): : 803 - 811
  • [5] Strategies for training optical neural networks
    Qipeng Yang
    Bowen Bai
    Weiwei Hu
    Xingjun Wang
    National Science Open, 2022, 1 (03) : 7 - 11
  • [6] Investigating Random Variations of the Forward-Forward Algorithm for Training Neural Networks
    Giampaolo, Fabio
    Izzo, Stefano
    Prezioso, Edoardo
    Piccialli, Francesco
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [7] Universal approximation of fully complex feed-forward neural networks
    Kim, T
    Adali, H
    2002 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-IV, PROCEEDINGS, 2002, : 973 - 976
  • [8] TRAINING FULLY RECURRENT NEURAL NETWORKS ON A RING TRANSPUTER ARRAY
    KECHRIOTIS, G
    MANOLAKOS, ES
    MICROPROCESSORS AND MICROSYSTEMS, 1994, 18 (01) : 5 - 11
  • [9] Training Fully Connected Neural Networks is ∃R-Complete
    Bertschinger, Daniel
    Hertrich, Christoph
    Jungeblut, Paul
    Miltzow, Tillmann
    Weber, Simon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Unsupervised Pre-training for Fully Convolutional Neural Networks
    Wiehman, Stiaan
    Kroon, Steve
    de Villiers, Hendrik
    2016 PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA AND ROBOTICS AND MECHATRONICS INTERNATIONAL CONFERENCE (PRASA-ROBMECH), 2016,