Pruning for Power: Optimizing Energy Efficiency in IoT with Neural Network Pruning

被引:0
|
作者
Widmann, Thomas [1 ]
Merkle, Florian [1 ]
Nocker, Martin [1 ]
Schoettle, Pascal [1 ]
机构
[1] MCI Management Ctr Innsbruck, Innsbruck, Austria
基金
奥地利科学基金会;
关键词
D O I
10.1007/978-3-031-34204-2_22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Internet of Things (IoT) has rapidly emerged as a crucial driver of the digital economy, generating massive amounts of data. Machine learning (ML) is an important technology to extract insights from the data generated by IoT devices. Deploying ML on low-power devices such as microcontroller units (MCUs) improves data protection, reduces bandwidth, and enables on-device data processing. However, the requirements of ML algorithms exceed the processing power, memory, and energy consumption capabilities of these devices. One solution to adapt ML networks to the limited capacities of MCUs is network pruning, the process of removing unnecessary connections or neurons from a neural network. In this work, we investigate the effect of unstructured and structured pruning methods on energy consumption. A series of experiments is conducted using a Raspberry Pi Pico to classify the FashionMNIST dataset with a LeNet-5-like convolutional neural network while applying unstructured magnitude and structured APoZ pruning approaches with various model compression rates from two to 64. We find that unstructured pruning out of the box has no effect on energy consumption, while structured pruning reduces energy consumption with increasing model compression. When structured pruning is applied to remove 75% of the model parameters, inference consumes 59.06% less energy, while the accuracy declines by 3.01 %. We further develop an adaption of the TensorFlow Lite framework that realizes the theoretical improvements for unstructured pruning, reducing the energy consumption by 37.59% with a decrease of only 1.58% in accuracy when 75% of the parameters are removed. Our results show that both approaches are feasible to significantly reduce the energy consumption of MCUs, leading to various possible sweet spots within the trade-off between accuracy and energy consumption.
引用
收藏
页码:251 / 263
页数:13
相关论文
共 50 条
  • [1] Optimizing Neural Network Efficiency with Hybrid Magnitude-Based and Node Pruning for Energy-efficient Computing in IoT
    Uddin, Mohammad Helal
    Baidya, Sabur
    PROCEEDINGS 8TH ACM/IEEE CONFERENCE ON INTERNET OF THINGS DESIGN AND IMPLEMENTATION, IOTDI 2023, 2023, : 474 - 476
  • [2] ScoringNet: A Neural Network Based Pruning Criteria for Structured Pruning
    Wang S.
    Zhang Z.
    Scientific Programming, 2023, 2023
  • [3] Recall Distortion in Neural Network Pruning and the Undecayed Pruning Algorithm
    Good, Aidan
    Lin, Jiaqi
    Yu, Xin
    Sieg, Hannah
    Ferguson, Mikey
    Zhe, Shandian
    Wieczorek, Jerzy
    Serra, Thiago
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [4] Federated Pruning: Improving Neural Network Efficiency with Federated Learning
    Lin, Rongmei
    Xiao, Yonghui
    Yang, Tien-Ju
    Zhao, Ding
    Xiong, Li
    Motta, Giovanni
    Beaufays, Francoise
    INTERSPEECH 2022, 2022, : 1701 - 1705
  • [5] Pruning by explaining: A novel criterion for deep neural network pruning
    Yeom, Seul-Ki
    Seegerer, Philipp
    Lapuschkin, Sebastian
    Binder, Alexander
    Wiedemann, Simon
    Mueller, Klaus-Robert
    Samek, Wojciech
    PATTERN RECOGNITION, 2021, 115
  • [6] A Novel Pruning Algorithm for Optimizing Feedforward Neural Network of Classification Problems
    Augasta, M. Gethsiyal
    Kathirvalavakumar, T.
    NEURAL PROCESSING LETTERS, 2011, 34 (03) : 241 - 258
  • [7] A Novel Pruning Algorithm for Optimizing Feedforward Neural Network of Classification Problems
    M. Gethsiyal Augasta
    T. Kathirvalavakumar
    Neural Processing Letters, 2011, 34 : 241 - 258
  • [8] Variational Convolutional Neural Network Pruning
    Zhao, Chenglong
    Ni, Bingbing
    Zhang, Jian
    Zhao, Qiwei
    Zhang, Wenjun
    Tian, Qi
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2775 - 2784
  • [9] Dirichlet Pruning for Neural Network Compression
    Adamczewski, Kamil
    Park, Mijung
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [10] Convolutional Neural Network Pruning: A Survey
    Xu, Sheng
    Huang, Anran
    Chen, Lei
    Zhang, Baochang
    PROCEEDINGS OF THE 39TH CHINESE CONTROL CONFERENCE, 2020, : 7458 - 7463