REDS: Resource-Efficient Deep Subnetworks for Dynamic Resource Constraints

Francesco Corti, Balz Maag, Joachim Schauer, Ulrich Pferschy, Olga Saukh

Publikation: ArbeitspapierPreprint

Abstract

Deep models deployed on edge devices frequently encounter resource variability, which arises from fluctuating energy levels, timing constraints, or prioritization of other critical tasks within the system. State-of-the-art machine learning pipelines generate resource-agnostic models, not capable to adapt at runtime. In this work we introduce Resource-Efficient Deep Subnetworks (REDS) to tackle model adaptation to variable resources. In contrast to the state-of-the-art, REDS use structured sparsity constructively by exploiting permutation invariance of neurons, which allows for hardware-specific optimizations. Specifically, REDS achieve computational efficiency by (1) skipping sequential computational blocks identified by a novel iterative knapsack optimizer, and (2) leveraging simple math to re-arrange the order of operations in REDS computational graph to take advantage of the data cache. REDS support conventional deep networks frequently deployed on the edge and provide computational benefits even for small and simple networks. We evaluate REDS on six benchmark architectures trained on the Google Speech Commands, FMNIST and CIFAR10 datasets, and test on four off-the-shelf mobile and embedded hardware platforms. We provide a theoretical result and empirical evidence for REDS outstanding performance in terms of submodels’ test set accuracy, and demonstrate an adaptation time in response to dynamic resource constraints of under 40μs, utilizing a 2-layer fully-connected network on Arduino Nano 33 BLE Sense.
Originalspracheenglisch
Seiten1-26
Seitenumfang26
PublikationsstatusEingereicht - 30 Okt. 2023

Fingerprint

Untersuchen Sie die Forschungsthemen von „REDS: Resource-Efficient Deep Subnetworks for Dynamic Resource Constraints“. Zusammen bilden sie einen einzigartigen Fingerprint.

Dieses zitieren