IEEE Access (Jan 2022)

Attention-Based Neural Bag-of-Features Learning for Sequence Data

  • Dat Thanh Tran,
  • Nikolaos Passalis,
  • Anastasios Tefas,
  • Moncef Gabbouj,
  • Alexandros Iosifidis

DOI
https://doi.org/10.1109/ACCESS.2022.3169776
Journal volume & issue
Vol. 10
pp. 45542 – 45552

Abstract

Read online

In this paper, we propose 2D-Attention (2DA), a generic attention formulation for sequence data, which acts as a complementary computation block that can detect and focus on relevant sources of information for the given learning objective. The proposed attention module is incorporated into the recently proposed Neural Bag of Feature (NBoF) model to enhance its learning capacity. Since 2DA acts as a plug-in layer, injecting it into different computation stages of the NBoF model results in different 2DA-NBoF architectures, each of which possesses a unique interpretation. We conducted extensive experiments in financial forecasting, audio analysis as well as medical diagnosis problems to benchmark the proposed formulations in comparison with existing methods, including the widely used Gated Recurrent Units. Our empirical analysis shows that the proposed attention formulations can not only improve performances of NBoF models but also make them resilient to noisy data.

Keywords