ICLR 2018
Skip to yearly menu bar Skip to main content


Workshop

Weightless: Lossy weight encoding for deep neural network compression

Brandon Reagen · Udit Gupta · Robert Adolf · Michael Mitzenmacher · Alexander Rush · Alexander Rush · Gu-Yeon Wei

East Meeting Level 8 + 15 #23

The large memory requirements of deep neural networks limit their deployment and adoption on many devices. Model compression methods effectively reduce the memory requirements of these models, usually through applying transformations such as weight pruning or quantization. In this paper, we present a novel scheme for lossy weight encoding which complements conventional compression techniques. The encoding is based on the Bloomier filter, a probabilistic data structure that can save space at the cost of introducing random errors. Leveraging the ability of neural networks to tolerate these imperfections and by re-training around the errors, the proposed technique, Weightless, can compress DNN weights by up to 496× with the same model accuracy. This results in up to a 1.51× improvement over the state-of-the-art.

Live content is unavailable. Log in and register to view live content