HADS: Hardware-Aware Deep Subnetworks

Francesco Corti*, Balz Maag, Joachim Schauer, Ulrich Pferschy, Olga Saukh

*Corresponding author for this work

Research output: Contribution to conferencePaperpeer-review

Abstract

We propose Hardware-Aware Deep Subnetworks (HADS) to tackle model adapta-
tion to dynamic resource contraints. In contrast to the state-of-the-art, HADS use
structured sparsity constructively by exploiting permutation invariance of neurons,
which allows for hardware-specific optimizations. HADS achieve computational
efficiency by skipping sequential computational blocks identified by a novel iter-
ative knapsack optimizer. HADS support conventional deep networks frequently
deployed on low-resource edge devices and provide computational benefits even
for small and simple networks. We evaluate HADS on six benchmark architec-
tures trained on the GOOGLE SPEECH COMMANDS, FMNIST and CIFAR10
datasets, and test on four off-the-shelf mobile and embedded hardware platforms.
We provide a theoretical result and empirical evidence for HADS outstanding per-
formance in terms of submodels’ test set accuracy, and demonstrate an adaptation
time in response to dynamic resource constraints of under 40μs, utilizing a 2-layer
fully-connected network on Arduino Nano 33 BLE Sense.
Original languageEnglish
Number of pages10
Publication statusAccepted/In press - 7 May 2024
EventInternational Conference on Learning Representations: ICLR 2024 - Messe Wien Exhibition and Congress Center, Wien, Austria
Duration: 7 May 202411 Jul 2024
https://iclr.cc/

Conference

ConferenceInternational Conference on Learning Representations
Abbreviated titleICLR
Country/TerritoryAustria
CityWien
Period7/05/2411/07/24
Internet address

Fingerprint

Dive into the research topics of 'HADS: Hardware-Aware Deep Subnetworks'. Together they form a unique fingerprint.

Cite this