Batteryless edge devices are extremely resource-constrained compared to traditional mobile platforms. Existing tiny deep neural network (DNN) inference solutions are problematic due to their slow and resource-intensive nature, rendering them unsuitable for batteryless edge devices. To address this problem, we propose a new approach to embedded intelligence, called Fast-Inf, which achieves extremely lightweight computation and minimal latency. Fast-Inf uses binary tree-based neural networks that are ultra-fast and energy-efficient due to their logarithmic time complexity. Additionally, Fast-Inf models can skip the leaf nodes when necessary, further minimizing latency without requiring any modifications to the model or retraining. Moreover, Fast-Inf models have significantly lower backup and runtime memory overhead. Our experiments on an MSP430FR5994 platform showed that Fast-Inf can achieve ultra-fast and energy-efficient inference (up to 700× speedup and reduced energy) compared to a conventional DNN.
Fast-Inf: Ultra-Fast Embedded Intelligence on the Batteryless Edge / Custode, Leonardo Lucio; Farina, Pietro; Yildiz, Eren; Kilic, Renan Beran; Yildirim, Kasim Sinan; Iacca, Giovanni. - (2024), pp. 239-252. (Intervento presentato al convegno SENSYS tenutosi a Hangzhou China nel 4th –7th November 2024) [10.1145/3666025.3699335].
Fast-Inf: Ultra-Fast Embedded Intelligence on the Batteryless Edge
Custode, Leonardo Lucio;Farina, Pietro;Yildiz, Eren;Kilic, Renan Beran;Yildirim, Kasim Sinan;Iacca, Giovanni
2024-01-01
Abstract
Batteryless edge devices are extremely resource-constrained compared to traditional mobile platforms. Existing tiny deep neural network (DNN) inference solutions are problematic due to their slow and resource-intensive nature, rendering them unsuitable for batteryless edge devices. To address this problem, we propose a new approach to embedded intelligence, called Fast-Inf, which achieves extremely lightweight computation and minimal latency. Fast-Inf uses binary tree-based neural networks that are ultra-fast and energy-efficient due to their logarithmic time complexity. Additionally, Fast-Inf models can skip the leaf nodes when necessary, further minimizing latency without requiring any modifications to the model or retraining. Moreover, Fast-Inf models have significantly lower backup and runtime memory overhead. Our experiments on an MSP430FR5994 platform showed that Fast-Inf can achieve ultra-fast and energy-efficient inference (up to 700× speedup and reduced energy) compared to a conventional DNN.File | Dimensione | Formato | |
---|---|---|---|
3666025.3699335.pdf
accesso aperto
Tipologia:
Versione editoriale (Publisher’s layout)
Licenza:
Creative commons
Dimensione
1.18 MB
Formato
Adobe PDF
|
1.18 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione