Mathematics (Jan 2023)

Machine-Learning Methods on Noisy and Sparse Data

  • Konstantinos Poulinakis,
  • Dimitris Drikakis,
  • Ioannis W. Kokkinakis,
  • Stephen Michael Spottswood

DOI
https://doi.org/10.3390/math11010236
Journal volume & issue
Vol. 11, no. 1
p. 236

Abstract

Read online

Experimental and computational data and field data obtained from measurements are often sparse and noisy. Consequently, interpolating unknown functions under these restrictions to provide accurate predictions is very challenging. This study compares machine-learning methods and cubic splines on the sparsity of training data they can handle, especially when training samples are noisy. We compare deviation from a true function f using the mean square error, signal-to-noise ratio and the Pearson R2 coefficient. We show that, given very sparse data, cubic splines constitute a more precise interpolation method than deep neural networks and multivariate adaptive regression splines. In contrast, machine-learning models are robust to noise and can outperform splines after a training data threshold is met. Our study aims to provide a general framework for interpolating one-dimensional signals, often the result of complex scientific simulations or laboratory experiments.

Keywords