LegoNet: Memory Footprint Reduction Through Block Weight Clustering
arXiv:2603.06606v1 Announce Type: new
Abstract: As the need for neural network-based applications to become more accurate and powerful grows, so too does their size and memory footprint. With embedded devices, whose cache and RAM are limited, this growth hinders their ability to leverage state-of-the-art neural network architectures. In this work, we propose textbf{LegoNet}, a compression technique that textbf{constructs blocks of weights of the entire model regardless of layer type} and clusters these induced blocks. Using blocks instead of individual values to cluster the weights, we were able to compress ResNet-50 trained for Cifar-10 and ImageNet with only 32 4×4 blocks, compressing the memory footprint by over a factor of textbf{64x without having to remove any weights} or changing the architecture and textbf{no loss to accuracy}, nor retraining or any data, and show how to find an arrangement of 16 4×4 blocks that gives a compression ratio of textbf{128x with less than 3% accuracy loss}. This was all achieved with textbf{no need for (re)training or fine-tuning}.