Many embedded applications have strict energy, memory, and time constraints, making neural network (NN) inference particularly challenging. Recently, a novel NN architecture called Fast Feedforward Networks (FFFs) has been proposed to achieve inference with extremely lightweight computational demands and minimal latency. Yet, the memory footprint of such NNs remains a challenge. In this paper, we attempt to overcome this challenge by using a weight-sharing technique, called weight virtualization, proposing different virtualization methods that take advantage of the peculiarities of the FFFs’ tree-based architecture. We further optimize the model’s size (resulting from the virtualization configuration) and performance via multi-objective evolutionary optimization based on NSGA-II. Our experiments (https://github.com/DIOL-UniTN/MOE-VFFF) show that, in different benchmarks, leaf virtualization can reduce the memory footprint by up to 13x with negligible accuracy loss.
Many embedded applications have strict energy, memory, and time constraints, making neural network (NN) inference particularly challenging. Recently, a novel NN architecture called Fast Feedforward Networks (FFFs) has been proposed to achieve inference with extremely lightweight computational demands and minimal latency. Yet, the memory footprint of such NNs remains a challenge. In this paper, we attempt to overcome this challenge by using a weight-sharing technique, called weight virtualization, proposing different virtualization methods that take advantage of the peculiarities of the FFFs’ tree-based architecture. We further optimize the model’s size (resulting from the virtualization configuration) and performance via multi-objective evolutionary optimization based on NSGA-II. Our experiments (https://github.com/DIOL-UniTN/MOE-VFFF) show that, in different benchmarks, leaf virtualization can reduce the memory footprint by up to 13x with negligible accuracy loss.
Multi-objective Evolutionary Optimization of Virtualized Fast Feedforward Networks / Kilic, Renan Beran; Yildirim, Kasim Sinan; Iacca, Giovanni. - 15612:(2025), pp. 270-286. ( 28th European Conference on Applications of Evolutionary Computation, EvoApplications 2025, held as part of EvoStar 2025 Trieste 23rd April-25th April 2025) [10.1007/978-3-031-90062-4_17].
Multi-objective Evolutionary Optimization of Virtualized Fast Feedforward Networks
Renan Beran Kilic;Kasim Sinan Yildirim;Giovanni Iacca
2025-01-01
Abstract
Many embedded applications have strict energy, memory, and time constraints, making neural network (NN) inference particularly challenging. Recently, a novel NN architecture called Fast Feedforward Networks (FFFs) has been proposed to achieve inference with extremely lightweight computational demands and minimal latency. Yet, the memory footprint of such NNs remains a challenge. In this paper, we attempt to overcome this challenge by using a weight-sharing technique, called weight virtualization, proposing different virtualization methods that take advantage of the peculiarities of the FFFs’ tree-based architecture. We further optimize the model’s size (resulting from the virtualization configuration) and performance via multi-objective evolutionary optimization based on NSGA-II. Our experiments (https://github.com/DIOL-UniTN/MOE-VFFF) show that, in different benchmarks, leaf virtualization can reduce the memory footprint by up to 13x with negligible accuracy loss.| File | Dimensione | Formato | |
|---|---|---|---|
|
paper_53.pdf
embargo fino al 17/04/2026
Tipologia:
Post-print referato (Refereed author’s manuscript)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
1.06 MB
Formato
Adobe PDF
|
1.06 MB | Adobe PDF | Visualizza/Apri |
|
Multi-objective Evolutionary Optimization of Virtualized Fast Feedforward Networks.pdf
Solo gestori archivio
Tipologia:
Versione editoriale (Publisher’s layout)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
1.82 MB
Formato
Adobe PDF
|
1.82 MB | Adobe PDF | Visualizza/Apri |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione



