Batteryless edge devices are extremely resource-constrained compared to traditional mobile platforms. Existing tiny deep neural network (DNN) inference solutions are problematic due to their slow and resource-intensive nature, rendering them unsuitable for batteryless edge devices. To address this problem, we propose a new approach to embedded intelligence, called Fast-Inf, which achieves extremely lightweight computation and minimal latency. Fast-Inf uses binary tree-based neural networks that are ultra-fast and energy-efficient due to their logarithmic time complexity. Additionally, Fast-Inf models can skip the leaf nodes when necessary, further minimizing latency without requiring any modifications to the model or retraining. Moreover, Fast-Inf models have significantly lower backup and runtime memory overhead. Our experiments on an MSP430FR5994 platform showed that Fast-Inf can achieve ultra-fast and energy-efficient inference (up to 700× speedup and reduced energy) compared to a conventional DNN.

Fast-Inf: Ultra-Fast Embedded Intelligence on the Batteryless Edge / Custode, Leonardo Lucio; Farina, Pietro; Yildiz, Eren; Kilic, Renan Beran; Yildirim, Kasim Sinan; Iacca, Giovanni. - (2024), pp. 239-252. (Intervento presentato al convegno SENSYS tenutosi a Hangzhou China nel 4th –7th November 2024) [10.1145/3666025.3699335].

Fast-Inf: Ultra-Fast Embedded Intelligence on the Batteryless Edge

Custode, Leonardo Lucio;Farina, Pietro;Yildiz, Eren;Kilic, Renan Beran;Yildirim, Kasim Sinan;Iacca, Giovanni
2024-01-01

Abstract

Batteryless edge devices are extremely resource-constrained compared to traditional mobile platforms. Existing tiny deep neural network (DNN) inference solutions are problematic due to their slow and resource-intensive nature, rendering them unsuitable for batteryless edge devices. To address this problem, we propose a new approach to embedded intelligence, called Fast-Inf, which achieves extremely lightweight computation and minimal latency. Fast-Inf uses binary tree-based neural networks that are ultra-fast and energy-efficient due to their logarithmic time complexity. Additionally, Fast-Inf models can skip the leaf nodes when necessary, further minimizing latency without requiring any modifications to the model or retraining. Moreover, Fast-Inf models have significantly lower backup and runtime memory overhead. Our experiments on an MSP430FR5994 platform showed that Fast-Inf can achieve ultra-fast and energy-efficient inference (up to 700× speedup and reduced energy) compared to a conventional DNN.
2024
SENSYS '24: Proceedings of the 22nd ACM Conference on Embedded Networked Sensor Systems
New York
ACM
979-8-4007-0697-4
Custode, Leonardo Lucio; Farina, Pietro; Yildiz, Eren; Kilic, Renan Beran; Yildirim, Kasim Sinan; Iacca, Giovanni
Fast-Inf: Ultra-Fast Embedded Intelligence on the Batteryless Edge / Custode, Leonardo Lucio; Farina, Pietro; Yildiz, Eren; Kilic, Renan Beran; Yildirim, Kasim Sinan; Iacca, Giovanni. - (2024), pp. 239-252. (Intervento presentato al convegno SENSYS tenutosi a Hangzhou China nel 4th –7th November 2024) [10.1145/3666025.3699335].
File in questo prodotto:
File Dimensione Formato  
3666025.3699335.pdf

accesso aperto

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Creative commons
Dimensione 1.18 MB
Formato Adobe PDF
1.18 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/437832
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact