Pruning for Power: Optimizing Energy Efficiency in IoT with Neural Network Pruning

被引:0
|
作者
Widmann, Thomas [1 ]
Merkle, Florian [1 ]
Nocker, Martin [1 ]
Schoettle, Pascal [1 ]
机构
[1] MCI Management Ctr Innsbruck, Innsbruck, Austria
基金
奥地利科学基金会;
关键词
D O I
10.1007/978-3-031-34204-2_22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Internet of Things (IoT) has rapidly emerged as a crucial driver of the digital economy, generating massive amounts of data. Machine learning (ML) is an important technology to extract insights from the data generated by IoT devices. Deploying ML on low-power devices such as microcontroller units (MCUs) improves data protection, reduces bandwidth, and enables on-device data processing. However, the requirements of ML algorithms exceed the processing power, memory, and energy consumption capabilities of these devices. One solution to adapt ML networks to the limited capacities of MCUs is network pruning, the process of removing unnecessary connections or neurons from a neural network. In this work, we investigate the effect of unstructured and structured pruning methods on energy consumption. A series of experiments is conducted using a Raspberry Pi Pico to classify the FashionMNIST dataset with a LeNet-5-like convolutional neural network while applying unstructured magnitude and structured APoZ pruning approaches with various model compression rates from two to 64. We find that unstructured pruning out of the box has no effect on energy consumption, while structured pruning reduces energy consumption with increasing model compression. When structured pruning is applied to remove 75% of the model parameters, inference consumes 59.06% less energy, while the accuracy declines by 3.01 %. We further develop an adaption of the TensorFlow Lite framework that realizes the theoretical improvements for unstructured pruning, reducing the energy consumption by 37.59% with a decrease of only 1.58% in accuracy when 75% of the parameters are removed. Our results show that both approaches are feasible to significantly reduce the energy consumption of MCUs, leading to various possible sweet spots within the trade-off between accuracy and energy consumption.
引用
收藏
页码:251 / 263
页数:13
相关论文
共 50 条
  • [21] Optimized Convolutional Neural Network at the IoT edge for image detection using pruning and quantization
    Soumyalatha Naveen
    Manjunath R Kounte
    Multimedia Tools and Applications, 2025, 84 (9) : 5435 - 5455
  • [22] Network pruning for energy saving in the Internet
    Cuomo, Francesca
    Cianfrani, Antonio
    Polverini, Marco
    Mangione, Daniele
    COMPUTER NETWORKS, 2012, 56 (10) : 2355 - 2367
  • [23] Power-Aware Pruning for Ultrafast, Energy-Efficient, and Accurate Optical Neural Network Design
    Hattori, Naoki
    Masuda, Yutaka
    Ishihara, Tohru
    Shinya, Akihiko
    Notomi, Masaya
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 1285 - 1290
  • [24] Intermittent-Aware Neural Network Pruning
    Lin, Chih-Chia
    Liu, Chia-Yin
    Yen, Chih-Hsuan
    Kuo, Tei-Wei
    Hsiu, Pi-Cheng
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [25] Pruning the deep neural network by similar function
    Liu, Hanqing
    Xin, Bo
    Mu, Senlin
    Zhu, Zhangqing
    2018 INTERNATIONAL SYMPOSIUM ON POWER ELECTRONICS AND CONTROL ENGINEERING (ISPECE 2018), 2019, 1187
  • [26] Quantisation and Pruning for Neural Network Compression and Regularisation
    Paupamah, Kimessha
    James, Steven
    Klein, Richard
    2020 INTERNATIONAL SAUPEC/ROBMECH/PRASA CONFERENCE, 2020, : 295 - 300
  • [27] Artificial Neural Network Pruning to Extract Knowledge
    Mirkes, Evgeny M.
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [28] ON THE ROLE OF STRUCTURED PRUNING FOR NEURAL NETWORK COMPRESSION
    Bragagnolo, Andrea
    Tartaglione, Enzo
    Fiandrotti, Attilio
    Grangetto, Marco
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 3527 - 3531
  • [29] Automated Pruning for Deep Neural Network Compression
    Manessi, Franco
    Rozza, Alessandro
    Bianco, Simone
    Napoletano, Paolo
    Schettini, Raimondo
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 657 - 664
  • [30] Effective pruning of neural network classifier ensembles
    Lazarevic, A
    Obradovic, Z
    IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 796 - 801