IEEE Access (Jan 2018)

BundleNet: Learning with Noisy Label via Sample Correlations

  • Chenghua Li,
  • Chunjie Zhang,
  • Kun Ding,
  • Gang Li,
  • Jian Cheng,
  • Hanqing Lu

DOI
https://doi.org/10.1109/ACCESS.2017.2782844
Journal volume & issue
Vol. 6
pp. 2367 – 2377

Abstract

Read online

Sequential patterns are important, because they can be exploited to improve the prediction accuracy of our classifiers. Sequential data, such as time series/video frames, and event data are becoming more and more ubiquitous in a wide spectrum of application scenarios especially in the background of large data and deep learning. However, large data sets used in training modern machine-learning models, such as deep neural networks, are often affected by label noise. Existing noisy learning approaches mainly focus on building an additional network to clean the noise or find a robust loss function. Few works tackle this problem by exploiting sample correlations. In this paper, we propose BundleNet, a framework of sequential structure (named bundle-module, see Fig. 1) for deep neural networks to handle the label noise. The bundle module naturally takes into account sample correlations by constructing bundles of samples class-by-class, and treats them as independent inputs. Moreover, we prove that the bundle-module performs a form of regularization, which is similar to dropout as regularization during training. The regularization effect endows the BundleNet with strong robustness to the label noise. Extensive experiments on public data sets prove that the proposed approach is effective and promising.

Keywords