Mirror Descent View for Neural Network Quantization

Thalaiyasingam Ajanthan, Kartik Gupta, Philip H.S. Torr, Richard Hartley, Puneet K. Dokania

    Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

    Abstract

    Quantizing large Neural Networks (NN) while maintaining the performance is highly desirable for resource-limited devices due to reduced memory and time complexity. It is usually formulated as a constrained optimization problem and optimized via a modified version of gradient descent. In this work, by interpreting the continuous parameters (unconstrained) as the dual of the quantized ones, we introduce a Mirror Descent (MD) framework (Bubeck (2015)) for NN quantization. Specifically, we provide conditions on the projections (i.e., mapping from continuous to quantized ones) which would enable us to derive valid mirror maps and in turn the respective MD updates. Furthermore, we present a numerically stable implementation of MD that requires storing an additional set of auxiliary variables (unconstrained), and show that it is strikingly analogous to the Straight Through Estimator (STE) based method which is typically viewed as a "trick" to avoid vanishing gradients issue. Our experiments on CIFAR-10/100, TinyImageNet, and ImageNet classification datasets with VGG-16, ResNet-18, and MobileNetV2 architectures show that our MD variants yield state-of-the-art performance.
    Original languageEnglish
    Title of host publicationProceedings of The 24th International Conference on Artificial Intelligence and Statistics, PMLR
    EditorsArindam Banerjee, Kenji Fukumizu
    Place of PublicationUnited States
    PublisherProceedings of Machine Learning Research
    Publication statusPublished - 2021
    Event24th International Conference on Artificial Intelligence and Statistics (AISTATS) 2021 - San Diego, California, USA
    Duration: 1 Jan 2021 → …

    Conference

    Conference24th International Conference on Artificial Intelligence and Statistics (AISTATS) 2021
    Period1/01/21 → …
    Other13 to 15 April 2021

    Fingerprint

    Dive into the research topics of 'Mirror Descent View for Neural Network Quantization'. Together they form a unique fingerprint.

    Cite this