Deep reinforcement learning with significant multiplications inference

被引:0
|
作者
Ivanov, Dmitry A. [1 ,2 ]
Larionov, Denis A. [2 ,3 ]
Kiselev, Mikhail V. [2 ,3 ]
Dylov, Dmitry V. [4 ,5 ]
机构
[1] Lomonosov Moscow State Univ, GSP 1,Leninskie Gory, Moscow 119991, Russia
[2] Cifrum, 3 Kholodilnyy per, Moscow 115191, Russia
[3] Chuvash State Univ, 15 Moskovsky pr, Cheboksary 428015, Chuvash, Russia
[4] Skolkovo Inst Sci & Technol, 30 1 Bolshoi blvd, Moscow 121205, Russia
[5] Artificial Intelligence Res Inst, 32 1 Kutuzovsky pr, Moscow 121170, Russia
基金
俄罗斯基础研究基金会;
关键词
D O I
10.1038/s41598-023-47245-y
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
We propose a sparse computation method for optimizing the inference of neural networks in reinforcement learning (RL) tasks. Motivated by the processing abilities of the brain, this method combines simple neural network pruning with a delta-network algorithm to account for the input data correlations. The former mimics neuroplasticity by eliminating inefficient connections; the latter makes it possible to update neuron states only when their changes exceed a certain threshold. This combination significantly reduces the number of multiplications during the neural network inference for fast neuromorphic computing. We tested the approach in popular deep RL tasks, yielding up to a 100-fold reduction in the number of required multiplications without substantial performance loss (sometimes, the performance even improved).
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Learning to Drive with Deep Reinforcement Learning
    Chukamphaeng, Nut
    Pasupa, Kitsuchart
    Antenreiter, Martin
    Auer, Peter
    2021 13TH INTERNATIONAL CONFERENCE ON KNOWLEDGE AND SMART TECHNOLOGY (KST-2021), 2021, : 147 - 152
  • [32] Deep Learning Type Inference
    Hellendoorn, Vincent J.
    Bird, Christian
    Barr, Earl T.
    Allamanis, Miltiadis
    ESEC/FSE'18: PROCEEDINGS OF THE 2018 26TH ACM JOINT MEETING ON EUROPEAN SOFTWARE ENGINEERING CONFERENCE AND SYMPOSIUM ON THE FOUNDATIONS OF SOFTWARE ENGINEERING, 2018, : 152 - 162
  • [33] A Survey on Reinforcement Learning and Deep Reinforcement Learning for Recommender Systems
    Rezaei, Mehrdad
    Tabrizi, Nasseh
    DEEP LEARNING THEORY AND APPLICATIONS, DELTA 2023, 2023, 1875 : 385 - 402
  • [34] Accuracy-Guaranteed Collaborative DNN Inference in Industrial IoT via Deep Reinforcement Learning
    Wu, Wen
    Yang, Peng
    Zhang, Weiting
    Zhou, Conghao
    Shen, Xuemin
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (07) : 4988 - 4998
  • [35] POMDP inference and robust solution via deep reinforcement learning: an application to railway optimal maintenance
    Arcieri, Giacomo
    Hoelzl, Cyprien
    Schwery, Oliver
    Straub, Daniel
    Papakonstantinou, Konstantinos G.
    Chatzi, Eleni
    MACHINE LEARNING, 2024, 113 (10) : 7967 - 7995
  • [36] Co-Adaptation of Algorithmic and Implementational Innovations in Inference-based Deep Reinforcement Learning
    Furuta, Hiroki
    Kozuno, Tadashi
    Matsushima, Tatsuya
    Matsuo, Yutaka
    Gu, Shixiang Shane
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [37] Deep Reinforcement Learning-based Text Anonymization against Private-Attribute Inference
    Mosallanezhad, Ahmadreza
    Beigi, Ghazaleh
    Liu, Huan
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 2360 - 2369
  • [38] Deep Reinforcement Learning for Containerized Edge Intelligence Inference Request Processing in IoT Edge Computing
    Nkenyereye, Lionel
    Baeg, Kang-Jun
    Chung, Wan-Young
    IEEE TRANSACTIONS ON SERVICES COMPUTING, 2023, 16 (06) : 4328 - 4344
  • [39] Guiding inference through relational reinforcement learning
    Asgharbeygi, N
    Nejati, N
    Langley, P
    Arai, S
    INDUCTIVE LOGIC PROGRAMMING, PROCEEDINGS, 2005, 3625 : 20 - 37
  • [40] Probabilistic inference for determining options in reinforcement learning
    Daniel, Christian
    van Hoof, Herke
    Peters, Jan
    Neumann, Gerhard
    MACHINE LEARNING, 2016, 104 (2-3) : 337 - 357