Optimization and inverse design of optical activation functions based on neural networks

被引:0
|
作者
Jia, Tao [1 ]
Jiang, Rui [1 ]
Fu, Ziling [1 ]
Xie, Zican [1 ]
Ding, Xin [1 ]
Wang, Zhi [1 ]
机构
[1] Beijing Jiaotong Univ, Sch Phys Sci & Engn, Inst Opt Informat, Key Lab Luminescence & Opt Informat,Minist Educ, Beijing 100044, Peoples R China
基金
国家重点研发计划;
关键词
Optical neural network; Optical nonlinear activation function; Mach-zehnder interferometer; Micro-ring resonator; PHOTONICS;
D O I
10.1016/j.optcom.2024.131370
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
The development of all-optical and electro-optical neural networks represents a rapidly growing field of research, with nonlinear activation functions serving as essential components of these systems. In this study, we employ an artificial neural network model to optimize the performance parameters of two systems based on Mach-Zehnder interferometers and micro-ring resonators. The results demonstrate that the optimized devices can accurately approximate several of the 14 activation functions (with a minimum root mean square error (RMSE) value of-33.1 dB), including Clipped ReLU, Sine, and Exponential. The optimized functions are also applied to an image recognition task using the Modified National Institute of Standards and Technology (MNIST) database, achieving maximum training and validation accuracies of 99.9% and 99.3% in simulation, respectively. Additionally, we introduce an inverse model to design the structural parameters of the coupling regions. Our approach significantly reduces the design time of the MZI-MRR activation function structure and theoretically demonstrates its feasibility and flexibility, providing a valuable example for the broader application of inverse design and optimization methods in optical neural network chips.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Microdisk modulator-assisted optical nonlinear activation functions for photonic neural networks
    Wang, Bin
    Yu, Weizhen
    Duan, Jinpeng
    Yang, Shuwen
    Zhao, Zhenyu
    Zheng, Shuang
    Zhang, Weifeng
    OPTICS COMMUNICATIONS, 2024, 553
  • [32] Quantum activation functions for quantum neural networks
    Marco Maronese
    Claudio Destri
    Enrico Prati
    Quantum Information Processing, 21
  • [33] HOLDER CONTINUOUS ACTIVATION FUNCTIONS IN NEURAL NETWORKS
    Tatar, Nasser-Eddine
    ADVANCES IN DIFFERENTIAL EQUATIONS AND CONTROL PROCESSES, 2015, 15 (02): : 93 - 106
  • [34] A Comparison of Activation Functions in Artificial Neural Networks
    Bircanoglu, Cenk
    Arica, Nafiz
    2018 26TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2018,
  • [35] Activation Functions and Their Characteristics in Deep Neural Networks
    Ding, Bin
    Qian, Huimin
    Zhou, Jun
    PROCEEDINGS OF THE 30TH CHINESE CONTROL AND DECISION CONFERENCE (2018 CCDC), 2018, : 1836 - 1841
  • [36] Deep Neural Networks with Multistate Activation Functions
    Cai, Chenghao
    Xu, Yanyan
    Ke, Dengfeng
    Su, Kaile
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2015, 2015
  • [37] Stochastic Neural Networks with Monotonic Activation Functions
    Ravanbakhsh, Siamak
    Poczos, Barnabas
    Schneider, Jeff
    Schuurmans, Dale
    Greiner, Russell
    ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 51, 2016, 51 : 809 - 818
  • [38] Quantum activation functions for quantum neural networks
    Maronese, Marco
    Destri, Claudio
    Prati, Enrico
    QUANTUM INFORMATION PROCESSING, 2022, 21 (04)
  • [39] Learning Activation Functions for Sparse Neural Networks
    Loni, Mohammad
    Mohan, Aditya
    Asadi, Mehdi
    Lindauer, Marius
    INTERNATIONAL CONFERENCE ON AUTOMATED MACHINE LEARNING, VOL 224, 2023, 224
  • [40] Fractional Adaptation of Activation Functions In Neural Networks
    Zamora Esquivel, Julio
    Cruz Vargas, Jesus Adan
    Lopez-Meyer, Paulo
    Cordourier Maruri, Hector Alfonso
    Camacho Perez, Jose Rodrigo
    Tickoo, Omesh
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7544 - 7550