Applied Sciences (Oct 2024)

Visual-Inertial Method for Localizing Aerial Vehicles in GNSS-Denied Environments

  • Andrea Tonini,
  • Mauro Castelli,
  • Jordan Steven Bates,
  • Nyi Nyi Nyan Lin,
  • Marco Painho

DOI
https://doi.org/10.3390/app14209493
Journal volume & issue
Vol. 14, no. 20
p. 9493

Abstract

Read online

Estimating the location of unmanned aerial vehicles (UAVs) within a global coordinate system can be achieved by correlating known world points with their corresponding image projections captured by the vehicle’s camera. Reducing the number of required world points may lower the computational requirements needed for such estimation. This paper introduces a novel method for determining the absolute position of aerial vehicles using only two known coordinate points that reduce the calculation complexity and, therefore, the computation time. The essential parameters for this calculation include the camera’s focal length, detector dimensions, and the Euler angles for Pitch and Roll. The Yaw angle is not required, which is beneficial because Yaw is more susceptible to inaccuracies due to environmental factors. The vehicle’s position is determined through a sequence of straightforward rigid transformations, eliminating the need for additional points or iterative processes for verification. The proposed method was tested using a Digital Elevation Model (DEM) created via LiDAR and 11 aerial images captured by a UAV. The results were compared against Global Navigation Satellite Systems (GNSSs) data and other common image pose estimation methodologies. While the available data did not permit precise error quantification, the method demonstrated performance comparable to GNSS-based approaches.

Keywords