Sensors (May 2024)

A Novel Frame-Selection Metric for Video Inpainting to Enhance Urban Feature Extraction

  • Yuhu Feng,
  • Jiahuan Zhang,
  • Guang Li,
  • Ren Togo,
  • Keisuke Maeda,
  • Takahiro Ogawa,
  • Miki Haseyama

DOI
https://doi.org/10.3390/s24103035
Journal volume & issue
Vol. 24, no. 10
p. 3035

Abstract

Read online

In our digitally driven society, advances in software and hardware to capture video data allow extensive gathering and analysis of large datasets. This has stimulated interest in extracting information from video data, such as buildings and urban streets, to enhance understanding of the environment. Urban buildings and streets, as essential parts of cities, carry valuable information relevant to daily life. Extracting features from these elements and integrating them with technologies such as VR and AR can contribute to more intelligent and personalized urban public services. Despite its potential benefits, collecting videos of urban environments introduces challenges because of the presence of dynamic objects. The varying shape of the target building in each frame necessitates careful selection to ensure the extraction of quality features. To address this problem, we propose a novel evaluation metric that considers the video-inpainting-restoration quality and the relevance of the target object, considering minimizing areas with cars, maximizing areas with the target building, and minimizing overlapping areas. This metric extends existing video-inpainting-evaluation metrics by considering the relevance of the target object and interconnectivity between objects. We conducted experiment to validate the proposed metrics using real-world datasets from Japanese cities Sapporo and Yokohama. The experiment results demonstrate feasibility of selecting video frames conducive to building feature extraction.

Keywords