EURASIP Journal on Advances in Signal Processing (Jun 2021)

An efficient pruning scheme of deep neural networks for Internet of Things applications

  • Chen Qi,
  • Shibo Shen,
  • Rongpeng Li,
  • Zhifeng Zhao,
  • Qing Liu,
  • Jing Liang,
  • Honggang Zhang

DOI
https://doi.org/10.1186/s13634-021-00744-4
Journal volume & issue
Vol. 2021, no. 1
pp. 1 – 21

Abstract

Read online

Abstract Nowadays, deep neural networks (DNNs) have been rapidly deployed to realize a number of functionalities like sensing, imaging, classification, recognition, etc. However, the computational-intensive requirement of DNNs makes it difficult to be applicable for resource-limited Internet of Things (IoT) devices. In this paper, we propose a novel pruning-based paradigm that aims to reduce the computational cost of DNNs, by uncovering a more compact structure and learning the effective weights therein, on the basis of not compromising the expressive capability of DNNs. In particular, our algorithm can achieve efficient end-to-end training that transfers a redundant neural network to a compact one with a specifically targeted compression rate directly. We comprehensively evaluate our approach on various representative benchmark datasets and compared with typical advanced convolutional neural network (CNN) architectures. The experimental results verify the superior performance and robust effectiveness of our scheme. For example, when pruning VGG on CIFAR-10, our proposed scheme is able to significantly reduce its FLOPs (floating-point operations) and number of parameters with a proportion of 76.2% and 94.1%, respectively, while still maintaining a satisfactory accuracy. To sum up, our scheme could facilitate the integration of DNNs into the common machine-learning-based IoT framework and establish distributed training of neural networks in both cloud and edge.

Keywords