Sensors (Jul 2022)

A Novel Deep-Learning Model Compression Based on Filter-Stripe Group Pruning and Its IoT Application

  • Ming Zhao,
  • Xindi Tong,
  • Weixian Wu,
  • Zhen Wang,
  • Bingxue Zhou,
  • Xiaodan Huang

DOI
https://doi.org/10.3390/s22155623
Journal volume & issue
Vol. 22, no. 15
p. 5623

Abstract

Read online

Nowadays, there is a tradeoff between the deep-learning module-compression ratio and the module accuracy. In this paper, a strategy for refining the pruning quantification and weights based on neural network filters is proposed. Firstly, filters in the neural network were refined into strip-like filter strips. Then, the evaluation of the filter strips was used to refine the partial importance of the filter, cut off the unimportant filter strips and reorganize the remaining filter strips. Finally, the training of the neural network after recombination was quantified to further compress the computational amount of the neural network. The results show that the method can significantly reduce the computational effort of the neural network and compress the number of parameters in the model. Based on experimental results on ResNet56, this method can reduce the number of parameters to 1/4 and the amount of calculation to 1/5, and the loss of model accuracy is only 0.01. On VGG16, the number of parameters is reduced to 1/14, the amount of calculation is reduced to 1/3, and the accuracy loss is 0.5%.

Keywords