Tunable Efficient Unitary Neural Networks (EUNN) and their application to RNNs

被引:0
|
作者
Li Jing [1 ]
Shen, Yichen [1 ]
Dubcek, Tena [1 ]
Peurifoy, John [1 ]
Skirlo, Scott [1 ]
LeCun, Yann [2 ]
Tegmark, Max [1 ]
Soljacic, Marin [1 ]
机构
[1] MIT, 77 Massachusetts Ave, Cambridge, MA 02139 USA
[2] New York Univ, Facebook AI Res, New York, NY 10003 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Using unitary (instead of general) matrices in artificial neural networks (ANNs) is a promising way to solve the gradient explosion/vanishing problem, as well as to enable ANNs to learn long-term correlations in the data. This approach appears particularly promising for Recurrent Neural Networks (RNNs). In this work, we present a new architecture for implementing an Efficient Unitary Neural Network (EUNNs); its main advantages can be summarized as follows. Firstly, the representation capacity of the unitary space in an EUNN is fully tunable, ranging from a subspace of SU(N) to the entire unitary space. Secondly, the computational complexity for training an EUNN is merely O(1) per parameter. Finally, we test the performance of EUNNs on the standard copying task, the pixel-permuted MNIST digit recognition benchmark as well as the Speech Prediction Test (TIMIT). We find that our architecture significantly outperforms both other state-of-the-art unitary RNNs and the LSTM architecture, in terms of the final performance and/or the wall-clock training speed. EUNNs are thus promising alternatives to RNNs and LSTMs for a wide variety of applications.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Efficient training of unitary optical neural networks
    Lu, Kunrun
    Guo, Xianxin
    OPTICS EXPRESS, 2023, 31 (24) : 39616 - 39623
  • [2] RNNs of RNNs: Recursive Construction of Stable Assemblies of Recurrent Neural Networks
    Kozachkov, Leo
    Ennis, Michaela
    Slotine, Jean-Jacques
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [3] Building Efficient Deep Neural Networks with Unitary Group Convolutions
    Zhao, Ritchie
    Hu, Yuwei
    Dotzel, Jordan
    De Sa, Christopher
    Zhang, Zhiru
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11295 - 11304
  • [4] NearUni: Near-Unitary Training for Efficient Optical Neural Networks
    Eldebiky, Amro
    Li, Bing
    Zhang, Grace Li
    2023 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2023,
  • [5] In-Depth Insights into the Application of Recurrent Neural Networks (RNNs) in Traffic Prediction: A Comprehensive Review
    He, Yuxin
    Huang, Ping
    Hong, Weihang
    Luo, Qin
    Li, Lishuai
    Tsui, Kwok-Leung
    ALGORITHMS, 2024, 17 (09)
  • [6] Recurrent neural networks (RNNs) learn the constitutive law of viscoelasticity
    Guang Chen
    Computational Mechanics, 2021, 67 : 1009 - 1019
  • [7] Recurrent neural networks (RNNs) learn the constitutive law of viscoelasticity
    Chen, Guang
    COMPUTATIONAL MECHANICS, 2021, 67 (03) : 1009 - 1019
  • [8] Deep Unitary Convolutional Neural Networks
    Chang, Hao-Yuan
    Wang, Kang L.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 170 - 181
  • [9] Unitary Evolution Recurrent Neural Networks
    Arjovsky, Martin
    Shah, Amar
    Bengio, Yoshua
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [10] Attention-Based Recurrent Neural Networks (RNNs) for Short Text Classification: An Application in Public Health Monitoring
    Edo-Osagie, Oduwa
    Lake, Iain
    Edeghere, Obaghe
    De La Iglesia, Beatriz
    ADVANCES IN COMPUTATIONAL INTELLIGENCE, IWANN 2019, PT I, 2019, 11506 : 895 - 911