Pruning for Power: Optimizing Energy Efficiency in IoT with Neural Network Pruning

被引:0
|
作者
Widmann, Thomas [1 ]
Merkle, Florian [1 ]
Nocker, Martin [1 ]
Schoettle, Pascal [1 ]
机构
[1] MCI Management Ctr Innsbruck, Innsbruck, Austria
基金
奥地利科学基金会;
关键词
D O I
10.1007/978-3-031-34204-2_22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Internet of Things (IoT) has rapidly emerged as a crucial driver of the digital economy, generating massive amounts of data. Machine learning (ML) is an important technology to extract insights from the data generated by IoT devices. Deploying ML on low-power devices such as microcontroller units (MCUs) improves data protection, reduces bandwidth, and enables on-device data processing. However, the requirements of ML algorithms exceed the processing power, memory, and energy consumption capabilities of these devices. One solution to adapt ML networks to the limited capacities of MCUs is network pruning, the process of removing unnecessary connections or neurons from a neural network. In this work, we investigate the effect of unstructured and structured pruning methods on energy consumption. A series of experiments is conducted using a Raspberry Pi Pico to classify the FashionMNIST dataset with a LeNet-5-like convolutional neural network while applying unstructured magnitude and structured APoZ pruning approaches with various model compression rates from two to 64. We find that unstructured pruning out of the box has no effect on energy consumption, while structured pruning reduces energy consumption with increasing model compression. When structured pruning is applied to remove 75% of the model parameters, inference consumes 59.06% less energy, while the accuracy declines by 3.01 %. We further develop an adaption of the TensorFlow Lite framework that realizes the theoretical improvements for unstructured pruning, reducing the energy consumption by 37.59% with a decrease of only 1.58% in accuracy when 75% of the parameters are removed. Our results show that both approaches are feasible to significantly reduce the energy consumption of MCUs, leading to various possible sweet spots within the trade-off between accuracy and energy consumption.
引用
收藏
页码:251 / 263
页数:13
相关论文
共 50 条
  • [31] Pruning for a faster, more transparent neural network
    Whitworth, CC
    MEASUREMENT & CONTROL, 1998, 31 (04): : 110 - 112
  • [32] Neural network pruning based on input importance
    Hewahi, Nabil M.
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2019, 37 (02) : 2243 - 2252
  • [33] Rethinking the Pruning Criteria for Convolutional Neural Network
    Huang, Zhongzhan
    Shao, Wenqi
    Wang, Xinjiang
    Lin, Liang
    Luo, Ping
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [34] A pruning algorithm for training neural network ensembles
    Shahjahan, M
    Akhand, MAH
    Murase, K
    SICE 2003 ANNUAL CONFERENCE, VOLS 1-3, 2003, : 628 - 633
  • [35] Robust Neural Network Pruning by Cooperative Coevolution
    Wu, Jia-Liang
    Shang, Haopu
    Hong, Wenjing
    Qian, Chao
    PARALLEL PROBLEM SOLVING FROM NATURE - PPSN XVII, PPSN 2022, PT I, 2022, 13398 : 459 - 473
  • [36] Pruning Convolutional Neural Network with Distinctiveness Approach
    Li, Wenrui
    Plested, Jo
    NEURAL INFORMATION PROCESSING, ICONIP 2019, PT V, 2019, 1143 : 448 - 455
  • [37] Constructive and pruning methods for neural network design
    Costa, MA
    Braga, AP
    de Menezes, BR
    VII BRAZILIAN SYMPOSIUM ON NEURAL NETWORKS, PROCEEDINGS, 2002, : 49 - 54
  • [38] Overview of Deep Convolutional Neural Network Pruning
    Li, Guang
    Liu, Fang
    Xia, Yuping
    2020 INTERNATIONAL CONFERENCE ON IMAGE, VIDEO PROCESSING AND ARTIFICIAL INTELLIGENCE, 2020, 11584
  • [39] Specialized Neural Network Pruning for Boolean Abstractions
    Briscoe, Jarren
    Rague, Brian
    Feuz, Kyle
    Ball, Robert
    PROCEEDINGS OF THE 13TH INTERNATIONAL JOINT CONFERENCE ON KNOWLEDGE DISCOVERY, KNOWLEDGE ENGINEERING AND KNOWLEDGE MANAGEMENT (KEOD), VOL 2, 2021, : 178 - 185
  • [40] Neural network pruning using MV regularizer
    Nagata, T
    Kawata, A
    Yamada, K
    Nakano, R
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 1051 - 1055