International Journal of Computational Intelligence Systems (Apr 2023)

Multiscale Feature Extraction and Fusion of Image and Text in VQA

  • Siyu Lu,
  • Yueming Ding,
  • Mingzhe Liu,
  • Zhengtong Yin,
  • Lirong Yin,
  • Wenfeng Zheng

DOI
https://doi.org/10.1007/s44196-023-00233-6
Journal volume & issue
Vol. 16, no. 1
pp. 1 – 11

Abstract

Read online

Abstract The Visual Question Answering (VQA) system is the process of finding useful information from images related to the question to answer the question correctly. It can be widely used in the fields of visual assistance, automated security surveillance, and intelligent interaction between robots and humans. However, the accuracy of VQA has not been ideal, and the main difficulty in its research is that the image features cannot well represent the scene and object information, and the text information cannot be fully represented. This paper used multi-scale feature extraction and fusion methods in the image feature characterization and text information representation sections of the VQA system, respectively to improve its accuracy. Firstly, aiming at the image feature representation problem, multi-scale feature extraction and fusion method were adopted, and the image features output of different network layers were extracted by a pre-trained deep neural network, and the optimal scheme of feature fusion method was found through experiments. Secondly, for the representation of sentences, a multi-scale feature method was introduced to characterize and fuse the word-level, phrase-level, and sentence-level features of sentences. Finally, the VQA model was improved using the multi-scale feature extraction and fusion method. The results show that the addition of multi-scale feature extraction and fusion improves the accuracy of the VQA model.

Keywords