IEEE Access (Jan 2019)

Inductive Zero-Shot Image Annotation via Embedding Graph

  • Fangxin Wang,
  • Jie Liu,
  • Shuwu Zhang,
  • Guixuan Zhang,
  • Yuejun Li,
  • Fei Yuan

DOI
https://doi.org/10.1109/ACCESS.2019.2925383
Journal volume & issue
Vol. 7
pp. 107816 – 107830

Abstract

Read online

Conventional image annotation systems can only handle those images having labels within the exist library, but cannot recognize those novel labels. In order to learn new concepts, one has to gather large amount of labeled images and train the model from scratch. More importantly, it can come with a high price to collect those labeled images. For these reasons, we put forward a zero-shot image annotation model, to reduce the demand for the images with novel labels. In this paper, we focus on the two big challenges of zero-shot image annotation: polysemous words and a strong bias in the generalized zero-shot setting. For the first problem, instead of training on large corpus datasets as previous methods, we propose to adopt Node2Vec to obtain contextualized word embeddings, which can easily produce word vectors of the polysemous words. For the second problem, we alleviate the strong bias in two ways: on one hand, we utilize a model based on graph convolutional network (GCN) to make target images involved in the training process; on the other hand, we put forward a novel semantic coherent (SC) loss to capture the semantic relations of the source and target labels. The extensive experiments on NUSWIDE, COCO, IAPR TC-12, and Corel5k datasets show the superiority of the proposed model and the annotation performance get improved by 4%-6% comparing with state-of-the-art methods.

Keywords