Timezone: »
Poster
MetaQuant: Learning to Quantize by Learning to Penetrate Non-differentiable Quantization
Shangyu Chen · Wenya Wang · Sinno Jialin Pan
Tue Dec 10 10:45 AM -- 12:45 PM (PST) @ East Exhibition Hall B + C #115
Tremendous amount of parameters make deep neural networks impractical to be deployed for edge-device-based real-world applications due to the limit of computational power and storage space. Existing studies have made progress on learning quantized deep models to reduce model size and energy consumption, i.e. converting full-precision weights ($r$'s) into discrete values ($q$'s) in a supervised training manner. However, the training process for quantization is non-differentiable, which leads to either infinite or zero gradients ($g_r$) w.r.t. $r$. To address this problem, most training-based quantization methods use the gradient w.r.t. $q$ ($g_q$) with clipping to approximate $g_r$ by Straight-Through-Estimator (STE) or manually design their computation. However, these methods only heuristically make training-based quantization applicable, without further analysis on how the approximated gradients can assist training of a quantized network. In this paper, we propose to learn $g_r$ by a neural network. Specifically, a meta network is trained using $g_q$ and $r$ as inputs, and outputs $g_r$ for subsequent weight updates. The meta network is updated together with the original quantized network. Our proposed method alleviates the problem of non-differentiability, and can be trained in an end-to-end manner. Extensive experiments are conducted with CIFAR10/100 and ImageNet on various deep networks to demonstrate the advantage of our proposed method in terms of a faster convergence rate and better performance. Codes are released at: \texttt{https://github.com/csyhhu/MetaQuant}
Author Information
Shangyu Chen (Nanyang Technological University, Singapore)
Wenya Wang (Nanyang Technological University)
Sinno Jialin Pan (Nanyang Technological University, Singapore)
More from the Same Authors
-
2020 Poster: Storage Efficient and Dynamic Flexible Runtime Channel Pruning via Deep Reinforcement Learning »
Jianda Chen · Shangyu Chen · Sinno Jialin Pan -
2017 Poster: Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon »
Xin Dong · Shangyu Chen · Sinno Pan -
2015 Workshop: Transfer and Multi-Task Learning: Trends and New Perspectives »
Anastasia Pentina · Christoph Lampert · Sinno Jialin Pan · Mingsheng Long · Judy Hoffman · Baochen Sun · Kate Saenko -
2013 Workshop: New Directions in Transfer and Multi-Task: Learning Across Domains and Tasks »
Urun Dogan · Marius Kloft · Tatiana Tommasi · Francesco Orabona · Massimiliano Pontil · Sinno Jialin Pan · Shai Ben-David · Arthur Gretton · Fei Sha · Marco Signoretto · Rajhans Samdani · Yun-Qian Miao · Mohammad Gheshlaghi azar · Ruth Urner · Christoph Lampert · Jonathan How -
2009 Workshop: Transfer Learning for Structured Data »
Sinno Jialin Pan · Ivor W Tsang · Le Song · Karsten Borgwardt · Qiang Yang