pith. machine review for the scientific record. sign in

arxiv: 1802.05668 · v1 · submitted 2018-02-15 · 💻 cs.NE · cs.LG

Recognition: unknown

Model compression via distillation and quantization

Authors on Pith no claims yet
classification 💻 cs.NE cs.LG
keywords distillationquantizationteachercompressionnetworksquantizedaccuracyadvances
0
0 comments X
read the original abstract

Deep neural networks (DNNs) continue to make significant advances, solving tasks from image classification to translation or reinforcement learning. One aspect of the field receiving considerable attention is efficiently executing deep models in resource-constrained environments, such as mobile or embedded devices. This paper focuses on this problem, and proposes two new compression methods, which jointly leverage weight quantization and distillation of larger teacher networks into smaller student networks. The first method we propose is called quantized distillation and leverages distillation during the training process, by incorporating distillation loss, expressed with respect to the teacher, into the training of a student network whose weights are quantized to a limited set of levels. The second method, differentiable quantization, optimizes the location of quantization points through stochastic gradient descent, to better fit the behavior of the teacher model. We validate both methods through experiments on convolutional and recurrent architectures. We show that quantized shallow students can reach similar accuracy levels to full-precision teacher models, while providing order of magnitude compression, and inference speedup that is linear in the depth reduction. In sum, our results enable DNNs for resource-constrained environments to leverage architecture and accuracy advances developed on more powerful devices.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Deep Learning for Protein Complex Prediction and Design

    cs.LG 2026-05 unverdicted novelty 3.0

    Deep learning architectures tailored to protein hierarchy combined with sequence-space search algorithms are used to improve prediction of protein complex structures and to design new interacting sequences.