AI-based IoT applications relying on heavy-load deep learning algorithms like CNNs challenge IoT devices that are restricted in energy or processing capabilities. Edge computing offers an alternative by allowing the data to get offloaded to so-called edge servers with hardware more powerful than IoT devices and physically closer than the cloud. However, the increasing complexity of data and algorithms and diverse conditions make even powerful devices, such as those equipped with FPGAs, insufficient to cope with the current demands. In this case, optimizations in the algorithms, like pruning and early-exit, are mandatory to reduce the CNNs computational burden and speed up inference processing. With that in mind, we propose ExpOL, which combines the pruning and early-exit CNN optimizations in a system-level FPGA-based IoT-Edge design space exploration. Based on a user-defined multi-target optimization, ExpOL delivers designs tailored to specific application environments and user needs. When evaluated against state-of-the-art FPGA-based accelerators (either local or offloaded), designs produced by ExpOL are more power-efficient (by up to 2x) and process inferences at higher user quality of experience (by up to 12.5%).
%0 Conference Paper
%1 korol_isvlsi23
%A Korol, Guilherme
%A Jordan, Michael Guilherme
%A Rutzig, Mateus Beck
%A Castrillon, Jeronimo
%A Beck, Antonio Carlos Schneider
%B 2023 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)
%C Los Alamitos, CA, USA
%D 2023
%E IEEE,
%I IEEE Computer Society
%K DNN FPGA accelerator
%R 10.1109/ISVLSI59464.2023.10238644
%T Design Space Exploration for CNN Offloading to FPGAs at the Edge
%U https://ieeexplore.ieee.org/abstract/document/10238644
%X AI-based IoT applications relying on heavy-load deep learning algorithms like CNNs challenge IoT devices that are restricted in energy or processing capabilities. Edge computing offers an alternative by allowing the data to get offloaded to so-called edge servers with hardware more powerful than IoT devices and physically closer than the cloud. However, the increasing complexity of data and algorithms and diverse conditions make even powerful devices, such as those equipped with FPGAs, insufficient to cope with the current demands. In this case, optimizations in the algorithms, like pruning and early-exit, are mandatory to reduce the CNNs computational burden and speed up inference processing. With that in mind, we propose ExpOL, which combines the pruning and early-exit CNN optimizations in a system-level FPGA-based IoT-Edge design space exploration. Based on a user-defined multi-target optimization, ExpOL delivers designs tailored to specific application environments and user needs. When evaluated against state-of-the-art FPGA-based accelerators (either local or offloaded), designs produced by ExpOL are more power-efficient (by up to 2x) and process inferences at higher user quality of experience (by up to 12.5%).
@inproceedings{korol_isvlsi23,
abstract = {AI-based IoT applications relying on heavy-load deep learning algorithms like CNNs challenge IoT devices that are restricted in energy or processing capabilities. Edge computing offers an alternative by allowing the data to get offloaded to so-called edge servers with hardware more powerful than IoT devices and physically closer than the cloud. However, the increasing complexity of data and algorithms and diverse conditions make even powerful devices, such as those equipped with FPGAs, insufficient to cope with the current demands. In this case, optimizations in the algorithms, like pruning and early-exit, are mandatory to reduce the CNNs computational burden and speed up inference processing. With that in mind, we propose ExpOL, which combines the pruning and early-exit CNN optimizations in a system-level FPGA-based IoT-Edge design space exploration. Based on a user-defined multi-target optimization, ExpOL delivers designs tailored to specific application environments and user needs. When evaluated against state-of-the-art FPGA-based accelerators (either local or offloaded), designs produced by ExpOL are more power-efficient (by up to 2x) and process inferences at higher user quality of experience (by up to 12.5%).},
added-at = {2025-01-02T10:40:48.000+0100},
address = {Los Alamitos, CA, USA},
author = {Korol, Guilherme and Jordan, Michael Guilherme and Rutzig, Mateus Beck and Castrillon, Jeronimo and Beck, Antonio Carlos Schneider},
biburl = {https://puma.scadsai.uni-leipzig.de/bibtex/2af8b8e0d719a71db15fc8708b9cd28fa/joca354e},
booktitle = {2023 IEEE Computer Society Annual Symposium on VLSI (ISVLSI)},
doi = {10.1109/ISVLSI59464.2023.10238644},
editor = {IEEE},
interhash = {b7fdc203a3530c784d6b54668cf52e06},
intrahash = {af8b8e0d719a71db15fc8708b9cd28fa},
keywords = {DNN FPGA accelerator},
location = {Foz do Igua{\c{c}}u, Brazil},
month = jun,
organization = {IEEE},
publisher = {IEEE Computer Society},
timestamp = {2025-01-02T10:40:48.000+0100},
title = {Design Space Exploration for {CNN} Offloading to {FPGAs} at the Edge},
url = {https://ieeexplore.ieee.org/abstract/document/10238644},
year = 2023
}