Timezone: »
After training complex deep learning models, a common task is to compress the model to reduce compute and storage demands. When compressing, it is desirable to preserve the original model's per-example decisions (e.g., to go beyond top-1 accuracy or preserve robustness), maintain the network's structure, automatically determine per-layer compression levels, and eliminate the need for fine tuning. No existing compression methods simultaneously satisfy these criteria---we introduce a principled approach that does by leveraging interpolative decompositions. Our approach simultaneously selects and eliminates channels (analogously, neurons), then constructs an interpolation matrix that propagates a correction into the next layer, preserving the network's structure. Consequently, our method achieves good performance even without fine tuning and admits theoretical analysis. Our theoretical generalization bound for a one layer network lends itself naturally to a heuristic that allows our method to automatically choose per-layer sizes for deep networks. We demonstrate the efficacy of our approach with strong empirical performance on a variety of tasks, models, and datasets---from simple one-hidden-layer networks to deep networks on ImageNet.
Author Information
Jerry Chee (Cornell)
Megan Flynn (née Renz) (Cornell University)
Anil Damle (Cornell University)
Christopher De Sa (Cornell University)
More from the Same Authors
-
2022 Poster: Understanding Hyperdimensional Computing for Parallel Single-Pass Learning »
Tao Yu · Yichi Zhang · Zhiru Zhang · Christopher De Sa -
2022 Poster: Communication-efficient distributed eigenspace estimation with arbitrary node failures »
Vasileios Charisopoulos · Anil Damle -
2022 Poster: GraB: Finding Provably Better Data Permutations than Random Reshuffling »
Yucheng Lu · Wentao Guo · Christopher De Sa -
2022 Poster: From Gradient Flow on Population Loss to Learning with Stochastic Gradient Descent »
Christopher De Sa · Satyen Kale · Jason Lee · Ayush Sekhari · Karthik Sridharan -
2020 Poster: Fast Matrix Square Roots with Applications to Gaussian Processes and Bayesian Optimization »
Geoff Pleiss · Martin Jankowiak · David Eriksson · Anil Damle · Jacob Gardner -
2020 Poster: Entrywise convergence of iterative methods for eigenproblems »
Vasileios Charisopoulos · Austin Benson · Anil Damle