Skip to yearly menu bar Skip to main content


Poster
in
Workshop: 5th Workshop on practical ML for limited/low resource settings (PML4LRS) @ ICLR 2024

HADS: Hardware-Aware Deep Subnetworks

Francesco Corti · Balz Maag · Joachim Schauer · Ulrich Pferschy · Olga Saukh


Abstract: We propose Hardware-Aware Deep Subnetworks (HADS) to tackle model adaptation to dynamic resource contraints. In contrast to the state-of-the-art, HADS use structured sparsity constructively by exploiting permutation invariance of neurons, which allows for hardware-specific optimizations. HADS achieve computational efficiency by skipping sequential computational blocks identified by a novel iterative knapsack optimizer. HADS support conventional deep networks frequently deployed on low-resource edge devices and provide computational benefits even for small and simple networks. We evaluate HADS on six benchmark architectures trained on the Google Speech Commands, Fashion-MNIST and CIFAR10 datasets, and test on four off-the-shelf mobile and embedded hardware platforms. We provide a theoretical result and empirical evidence for HADS outstanding performance in terms of submodels' test set accuracy, and demonstrate an adaptation time in response to dynamic resource constraints of under 40μs, utilizing a 2-layer fully-connected network on Arduino Nano 33 BLE Sense.

Chat is not available.