Batteryless systems frequently face power failures, requiring extra runtime buffers to maintain inference progress and leaving only a memory space for storing ultra-tiny deep neural networks (DNNs). Besides, making these models responsive to stochastic energy harvesting dynamics during inference requires a balance between inference accuracy, latency, and energy overhead. Recent works on compression mostly focus on time and memory, but often ignore energy dynamics or significantly reduce the accuracy of pre-trained DNNs. Existing energy-adaptive inference works modify the architecture of pre-trained models and have significant memory overhead. Thus, energy-adaptive and accurate inference of pre-trained DNNs on batteryless devices with extreme memory constraints is more challenging than traditional microcontrollers. Wecombat these issues by proposing FreeML, a framework to optimize pre-trained DNN models for memory-efficient and energyadaptive inference on batteryless systems. FreeML comprises (1) a novel compression technique to reduce the model footprint and runtime memory requirements simultaneously, making them executable on extremely memory-constrained batteryless platforms; and (2) the first early exit mechanism that uses a single exit branch for all exit points to terminate inference at any time, making models energy-adaptive with minimal memory overhead. Our experiments showed that FreeML reduces the model sizes by up to 95×, supports adaptive inference with a 2.03−19.65× less memory overhead, and provides significant time and energy benefits with only a negligible accuracy drop compared to the state-of-the-art.
Memory-efficient Energy-adaptive Inference of Pre-Trained Models on Batteryless Embedded Systems / Farina, Pietro; Biswas, Subrata; Yıldız, Eren; Akhunov, Khakim; Ahmed, Saad; Islam, Bashima; Yildirim, Kasim Sinan. - (2024). (Intervento presentato al convegno International Conference on embedded Wireless Systems and Networks (EWSN) tenutosi a Abu Dhabi, UAE nel 10th-13th December 2024).
Memory-efficient Energy-adaptive Inference of Pre-Trained Models on Batteryless Embedded Systems
Pietro Farina;Eren Yıldız;Khakim Akhunov;Kasım Sinan Yıldırım
2024-01-01
Abstract
Batteryless systems frequently face power failures, requiring extra runtime buffers to maintain inference progress and leaving only a memory space for storing ultra-tiny deep neural networks (DNNs). Besides, making these models responsive to stochastic energy harvesting dynamics during inference requires a balance between inference accuracy, latency, and energy overhead. Recent works on compression mostly focus on time and memory, but often ignore energy dynamics or significantly reduce the accuracy of pre-trained DNNs. Existing energy-adaptive inference works modify the architecture of pre-trained models and have significant memory overhead. Thus, energy-adaptive and accurate inference of pre-trained DNNs on batteryless devices with extreme memory constraints is more challenging than traditional microcontrollers. Wecombat these issues by proposing FreeML, a framework to optimize pre-trained DNN models for memory-efficient and energyadaptive inference on batteryless systems. FreeML comprises (1) a novel compression technique to reduce the model footprint and runtime memory requirements simultaneously, making them executable on extremely memory-constrained batteryless platforms; and (2) the first early exit mechanism that uses a single exit branch for all exit points to terminate inference at any time, making models energy-adaptive with minimal memory overhead. Our experiments showed that FreeML reduces the model sizes by up to 95×, supports adaptive inference with a 2.03−19.65× less memory overhead, and provides significant time and energy benefits with only a negligible accuracy drop compared to the state-of-the-art.| File | Dimensione | Formato | |
|---|---|---|---|
|
ewsn24-final16 (1).pdf
accesso aperto
Tipologia:
Versione editoriale (Publisher’s layout)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
4.07 MB
Formato
Adobe PDF
|
4.07 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione



