IEEE Access (Jan 2025)

High-Fidelity Depth Map Reconstruction System With RGB-Guided Super Resolution CNN and Cross-Calibrated Chaos LiDAR

  • Yu-Chun Ding,
  • Chia-Yu Chang,
  • Pei-Rong Li,
  • Chao-Tsung Huang,
  • Yung-Chen Lin,
  • Tsung Chen,
  • Wei-Lun Lin,
  • Cheng-Ting Lee,
  • Fan-Yi Lin,
  • Yuan-Hao Huang

DOI
https://doi.org/10.1109/ACCESS.2025.3532621
Journal volume & issue
Vol. 13
pp. 19118 – 19131

Abstract

Read online

High-fidelity 3D models are essential for immersive virtual and augmented reality (VR/AR) applications. However, the performance of current 3D recording devices is limited in several scenarios, such as dim light environments, long-distance measurements, and large-scale objects. Therefore, their applicability to indoor scenes is hindered. In this work, we propose a depth map reconstruction system that integrates an RGB-guided depth map super-resolution convolutional neural network (CNN) into a stand-alone Chaos LiDAR depth sensor. This system provides highly accurate depth estimates in various scenarios, particularly for indoor scenes with dim lighting or long distances ranging from 4 m to 6 m. We address two design challenges to maximize the quality of the reconstructed depth map of the system. First, the misalignment across RGB-depth sensors is addressed using a two-stage calibration pipeline. Second, the lack of large-scale real-world LiDAR datasets is addressed by generating a large-scale synthetic dataset and adopting transfer learning. Experimental results show that our proposed system significantly outperforms the commercial RGB-D recording device RealSense D435i in terms of subjective visual perception, precision, and density of depth estimates, making it a promising solution for general indoor scene recording.

Keywords