IEEE Access (Jan 2019)

Temporal Spiking Recurrent Neural Network for Action Recognition

  • Wei Wang,
  • Siyuan Hao,
  • Yunchao Wei,
  • Shengtao Xiao,
  • Jiashi Feng,
  • Nicu Sebe

DOI
https://doi.org/10.1109/ACCESS.2019.2936604
Journal volume & issue
Vol. 7
pp. 117165 – 117175

Abstract

Read online

In this paper, we propose a novel temporal spiking recurrent neural network (TSRNN) to perform robust action recognition in videos. The proposed TSRNN employs a novel spiking architecture which utilizes the local discriminative features from high-confidence reliable frames as spiking signals. The conventional CNN-RNNs typically used for this problem treat all the frames equally important such that they are error-prone to noisy frames. The TSRNN solves this problem by employing a temporal pooling architecture which can help RNN select sparse and reliable frames and enhances its capability in modelling long-range temporal information. Besides, a message passing bridge is added between the spiking signals and the recurrent unit. In this way, the spiking signals can guide RNN to correct its long-term memory across multiple frames from contamination caused by noisy frames with distracting factors (e.g., occlusion, rapid scene transition). With these two novel components, TSRNN achieves competitive performance compared with the state-of-the-art CNN-RNN architectures on two large scale public benchmarks, UCF101 and HMDB51.

Keywords