Machine Learning and Knowledge Extraction (Oct 2022)

FeaSel-Net: A Recursive Feature Selection Callback in Neural Networks

  • Felix Fischer,
  • Alexander Birk,
  • Peter Somers,
  • Karsten Frenner,
  • Cristina Tarín,
  • Alois Herkommer

DOI
https://doi.org/10.3390/make4040049
Journal volume & issue
Vol. 4, no. 4
pp. 968 – 993

Abstract

Read online

Selecting only the relevant subsets from all gathered data has never been as challenging as it is in these times of big data and sensor fusion. Multiple complementary methods have emerged for the observation of similar phenomena; oftentimes, many of these techniques are superimposed in order to make the best possible decisions. A pathologist, for example, uses microscopic and spectroscopic techniques to discriminate between healthy and cancerous tissue. Especially in the field of spectroscopy in medicine, an immense number of frequencies are recorded and appropriately sized datasets are rarely acquired due to the time-intensive measurements and the lack of patients. In order to cope with the curse of dimensionality in machine learning, it is necessary to reduce the overhead from irrelevant or redundant features. In this article, we propose a feature selection callback algorithm (FeaSel-Net) that can be embedded in deep neural networks. It recursively prunes the input nodes after the optimizer in the neural network achieves satisfying results. We demonstrate the performance of the feature selection algorithm on different publicly available datasets and compare it to existing feature selection methods. Our algorithm combines the advantages of neural networks’ nonlinear learning ability and the embedding of the feature selection algorithm into the actual classifier optimization.

Keywords