Batteryless systems frequently face power failures, requiring extra runtime buffers to maintain inference progress and leaving only a memory space for storing ultra-tiny deep neural networks (DNNs). Besides, making these models responsive to stochastic energy harvesting dynamics during inference requires a balance between inference accuracy, latency, and energy overhead. Recent works on compression mostly focus on time and memory, but often ignore energy dynamics or significantly reduce the accuracy of pre-trained DNNs. Existing energy-adaptive inference works modify the architecture of pre-trained models and have significant memory overhead. Thus, energy-adaptive and accurate inference of pre-trained DNNs on batteryless devices with extreme memory constraints is more challenging than traditional microcontrollers. Wecombat these issues by proposing FreeML, a framework to optimize pre-trained DNN models for memory-efficient and energyadaptive inference on batteryless systems. FreeML comprises (1) a novel compression technique to reduce the model footprint and runtime memory requirements simultaneously, making them executable on extremely memory-constrained batteryless platforms; and (2) the first early exit mechanism that uses a single exit branch for all exit points to terminate inference at any time, making models energy-adaptive with minimal memory overhead. Our experiments showed that FreeML reduces the model sizes by up to 95×, supports adaptive inference with a 2.03−19.65× less memory overhead, and provides significant time and energy benefits with only a negligible accuracy drop compared to the state-of-the-art.

Memory-efficient Energy-adaptive Inference of Pre-Trained Models on Batteryless Embedded Systems / Farina, Pietro; Biswas, Subrata; Yıldız, Eren; Akhunov, Khakim; Ahmed, Saad; Islam, Bashima; Yildirim, Kasim Sinan. - (2024). (Intervento presentato al convegno International Conference on embedded Wireless Systems and Networks (EWSN) tenutosi a Abu Dhabi, UAE nel 10th-13th December 2024).

Memory-efficient Energy-adaptive Inference of Pre-Trained Models on Batteryless Embedded Systems

Pietro Farina;Eren Yıldız;Khakim Akhunov;Kasım Sinan Yıldırım
2024-01-01

Abstract

Batteryless systems frequently face power failures, requiring extra runtime buffers to maintain inference progress and leaving only a memory space for storing ultra-tiny deep neural networks (DNNs). Besides, making these models responsive to stochastic energy harvesting dynamics during inference requires a balance between inference accuracy, latency, and energy overhead. Recent works on compression mostly focus on time and memory, but often ignore energy dynamics or significantly reduce the accuracy of pre-trained DNNs. Existing energy-adaptive inference works modify the architecture of pre-trained models and have significant memory overhead. Thus, energy-adaptive and accurate inference of pre-trained DNNs on batteryless devices with extreme memory constraints is more challenging than traditional microcontrollers. Wecombat these issues by proposing FreeML, a framework to optimize pre-trained DNN models for memory-efficient and energyadaptive inference on batteryless systems. FreeML comprises (1) a novel compression technique to reduce the model footprint and runtime memory requirements simultaneously, making them executable on extremely memory-constrained batteryless platforms; and (2) the first early exit mechanism that uses a single exit branch for all exit points to terminate inference at any time, making models energy-adaptive with minimal memory overhead. Our experiments showed that FreeML reduces the model sizes by up to 95×, supports adaptive inference with a 2.03−19.65× less memory overhead, and provides significant time and energy benefits with only a negligible accuracy drop compared to the state-of-the-art.
2024
EWSN '24: Proceedings of the 2023 International Conference on embedded Wireless Systems and Networks
New York, NY, USA
Association for Computing Machinery
Farina, Pietro; Biswas, Subrata; Yıldız, Eren; Akhunov, Khakim; Ahmed, Saad; Islam, Bashima; Yildirim, Kasim Sinan
Memory-efficient Energy-adaptive Inference of Pre-Trained Models on Batteryless Embedded Systems / Farina, Pietro; Biswas, Subrata; Yıldız, Eren; Akhunov, Khakim; Ahmed, Saad; Islam, Bashima; Yildirim, Kasim Sinan. - (2024). (Intervento presentato al convegno International Conference on embedded Wireless Systems and Networks (EWSN) tenutosi a Abu Dhabi, UAE nel 10th-13th December 2024).
File in questo prodotto:
File Dimensione Formato  
ewsn24-final16 (1).pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 4.07 MB
Formato Adobe PDF
4.07 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/445210
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact