BMC Research Notes (Nov 2020)

Can pre-trained convolutional neural networks be directly used as a feature extractor for video-based neonatal sleep and wake classification?

  • Muhammad Awais,
  • Xi Long,
  • Bin Yin,
  • Chen Chen,
  • Saeed Akbarzadeh,
  • Saadullah Farooq Abbasi,
  • Muhammad Irfan,
  • Chunmei Lu,
  • Xinhua Wang,
  • Laishuan Wang,
  • Wei Chen

DOI
https://doi.org/10.1186/s13104-020-05343-4
Journal volume & issue
Vol. 13, no. 1
pp. 1 – 6

Abstract

Read online

Abstract Objective In this paper, we propose to evaluate the use of pre-trained convolutional neural networks (CNNs) as a features extractor followed by the Principal Component Analysis (PCA) to find the best discriminant features to perform classification using support vector machine (SVM) algorithm for neonatal sleep and wake states using Fluke® facial video frames. Using pre-trained CNNs as a feature extractor would hugely reduce the effort of collecting new neonatal data for training a neural network which could be computationally expensive. The features are extracted after fully connected layers (FCL’s), where we compare several pre-trained CNNs, e.g., VGG16, VGG19, InceptionV3, GoogLeNet, ResNet, and AlexNet. Results From around 2-h Fluke® video recording of seven neonates, we achieved a modest classification performance with an accuracy, sensitivity, and specificity of 65.3%, 69.8%, 61.0%, respectively with AlexNet using Fluke® (RGB) video frames. This indicates that using a pre-trained model as a feature extractor could not fully suffice for highly reliable sleep and wake classification in neonates. Therefore, in future work a dedicated neural network trained on neonatal data or a transfer learning approach is required.

Keywords