Robust Architecture-Agnostic and Noise Resilient Training of Photonic Deep Learning Models

被引:17
|
作者
Kirtas, Manos [1 ]
Passalis, Nikolaos [1 ]
Mourgias-Alexandris, George [2 ]
Dabos, George [2 ]
Pleros, Nikos [2 ]
Tefas, Anastasios [1 ]
机构
[1] Aristotle Univ Thessaloniki, Dept Informat, Deep Learning Grp, Computat Intelligence, Thessaloniki 54124, Greece
[2] Aristotle Univ Thessaloniki, Dept Informat, Wireless, Networks Grp,Photon Syst, Thessaloniki 54124, Greece
基金
欧盟地平线“2020”;
关键词
Photonics; Training; Neurons; Neuromorphics; Modulation; Adaptation models; Optical noise; Photonic deep learning; neural network initialization; constrains-aware training;
D O I
10.1109/TETCI.2022.3182765
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neuromorphic photonic accelerators for Deep Learning (DL) have increasingly gained attention over the recent years due to their ability for ultra fast matrix-based calculations and low power consumption providing a great potential for DL implementations to deal with a wide range of different applications. At the same time, physical properties of the optical components hinder their application since they introduce a number of limitations, such as easily saturated activation functions as well as the existence of various noise sources. As a result, photonic DL models are especially challenging to be trained and deployed, compared with regular DL models, since traditionally used methods do not take into account the aforementioned constraints. To overcome these limitations and motivated by the fact that the information lost in one layer cannot be easily recovered when gradient-descent based algorithms are employed, we propose a novel training method for photonic neuromorphic architectures that is capable of taking into account a wide range of limitations of the actual hardware, including noise sources and easily saturated activation mechanisms. Compared to existing works, the proposed method takes a more holistic view of the training process, focusing both on the initialization process, as well as on the actual weight updates. The effectiveness of the proposed method is demonstrated on a variety of different problems and photonic neural network (PNN) architectures, including a noisy photonic recurrent neural network evaluated on high-frequency time series forecasting and a deep photonic feed-forward setup consisting of a transmitter, noisy channel, and receiver, which is used as an intensity modulation/direct detection system (IM/DD).
引用
收藏
页码:140 / 149
页数:10
相关论文
共 50 条
  • [31] Addressing The False Negative Problem of Deep Learning MRI Reconstruction Models by Adversarial Attacks and Robust Training
    Cheng, Kaiyang
    Caliva, Francesco
    Shah, Rutwik
    Han, Misung
    Majumdar, Sharmila
    Pedoia, Valentina
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 121, 2020, 121 : 121 - 135
  • [32] Training Deep Models Faster with Robust, Approximate Importance Sampling
    Johnson, Tyler B.
    Guestrin, Carlos
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [33] Robust Adversarial Objects against Deep Learning Models
    Tsai, Tzungyu
    Yang, Kaichen
    Ho, Tsung-Yi
    Jin, Yier
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 954 - 962
  • [34] Addressing Vulnerability in Medical Deep Learning through Robust Training
    Martinez-Martinez, Josue
    Nabavi, Sheida
    2023 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI, 2023, : 341 - 342
  • [35] Distributed deep learning training using silicon photonic switched architectures
    Zhu, Ziyi
    Teh, Min Yee
    Wu, Zhenguo
    Glick, Madeleine Strom
    Yan, Shijia
    Hattink, Maarten
    Bergman, Keren
    APL PHOTONICS, 2022, 7 (03)
  • [36] Evolution and Role of Optimizers in Training Deep Learning Models
    XiaoHao Wen
    MengChu Zhou
    IEEE/CAA Journal of Automatica Sinica, 2024, 11 (10) : 2039 - 2042
  • [37] Evolution and Role of Optimizers in Training Deep Learning Models
    Wen, XiaoHao
    Zhou, MengChu
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2024, 11 (10) : 2039 - 2042
  • [38] Anderson Acceleration for Distributed Training of Deep Learning Models
    Pasini, Massimiliano Lupo
    Yin, Junqi
    Reshniak, Viktor
    Stoyanov, Miroslav K.
    SOUTHEASTCON 2022, 2022, : 289 - 295
  • [39] Pulsed Thermography Dataset for Training Deep Learning Models
    Wei, Ziang
    Osman, Ahmad
    Valeske, Bernd
    Maldague, Xavier
    APPLIED SCIENCES-BASEL, 2023, 13 (05):
  • [40] Bypassing Stationary Points in Training Deep Learning Models
    Jung, Jaeheun
    Lee, Donghun
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 18859 - 18871