IEEE Access (Jan 2020)

FOSNet: An End-to-End Trainable Deep Neural Network for Scene Recognition

  • Hongje Seong,
  • Junhyuk Hyun,
  • Euntai Kim

DOI
https://doi.org/10.1109/ACCESS.2020.2989863
Journal volume & issue
Vol. 8
pp. 82066 – 82077

Abstract

Read online

Scene recognition is a kind of image recognition problems which is aimed at predicting the category of the place at which the image is taken. In this paper, a new scene recognition method using the convolutional neural network (CNN) is proposed. The proposed method is based on the fusion of the object and the scene information in the given image and the CNN framework is named as FOS (fusion of object and scene) Net. To combine the object and the scene information effectively, a new fusion framework named CCG (correlative context gating) is proposed. In addition, a new loss named scene coherence loss (SCL) is developed to train the FOSNet and to improve the scene recognition performance. The proposed SCL is based on the idea that the scene class does not change all over the image. The proposed FOSNet was experimented with three most popular scene recognition datasets, and their state-of-the-art performance is obtained in two sets: 60.14% on Places 2 and 90.30% on MIT indoor 67. The second highest performance of 77.28% is obtained on SUN 397.

Keywords