IEEE Access (Jan 2019)

Pruning Blocks for CNN Compression and Acceleration via Online Ensemble Distillation

  • Zongyue Wang,
  • Shaohui Lin,
  • Jiao Xie,
  • Yangbin Lin

DOI
https://doi.org/10.1109/ACCESS.2019.2957203
Journal volume & issue
Vol. 7
pp. 175703 – 175716

Abstract

Read online

In this paper, we propose an online ensemble distillation (OED) method to automatically prune blocks/layers of a target network by transferring the knowledge from a strong teacher in an end-to-end manner. To accomplish this, we first introduce a soft mask to scale the output of each block in the target network and enforce the sparsity of the mask by sparsity regularization. Then, a strong teacher network is constructed online by replicating the same target networks and ensembling the discriminative features from each target as its new features. Cooperative learning between multiple target networks and the teacher network is further conducted in a closed-loop form, which improves their performance. To solve the optimization problem in an end-to-end manner, we employ the fast iterative shrinkage-thresholding algorithm to fast and reliably remove the redundant blocks, in which the corresponding soft masks are equal to zero. Compared to other structured pruning methods with iterative fine-tuning, the proposed OED is trained more efficiently in one training cycle. Extensive experiments demonstrate the effectiveness of OED, which can not only simultaneously compress and accelerate a variety of CNN architectures but also enhance the robustness of the pruned networks.

Keywords