Complex & Intelligent Systems (Mar 2023)

Differentiable channel pruning guided via attention mechanism: a novel neural network pruning approach

  • Hanjing Cheng,
  • Zidong Wang,
  • Lifeng Ma,
  • Zhihui Wei,
  • Fawaz E. Alsaadi,
  • Xiaohui Liu

DOI
https://doi.org/10.1007/s40747-023-01022-6
Journal volume & issue
Vol. 9, no. 5
pp. 5611 – 5624

Abstract

Read online

Abstract Neural network pruning offers great prospects for facilitating the deployment of deep neural networks on computational resource limited devices. Neural architecture search (NAS) provides an efficient way to automatically seek appropriate neural architecture design for compressed model. It is observed that, for existing NAS-based pruning methods, there is usually a lack of layer information when searching the optimal neural architecture. In this paper, we propose a new NAS approach, namely, differentiable channel pruning method guided via attention mechanism (DCP-A), where the adopted attention mechanism is able to provide layer information to guide the optimization of the pruning policy. The training process is differentiable with Gumbel-softmax sampling, while parameters are optimized under a two-stage training procedure. The neural network block with the shortcut is dedicatedly designed, which is of help to prune the network not only on its width but also on its depth. Extensive experiments are performed to verify the applicability and superiority of the proposed method. Detailed analysis with visualization of the pruned model architecture shows that our proposed DCP-A learns explainable pruning policies.

Keywords