IEEE Access (Jan 2022)

Nonlinear Depth Quantization Using Piecewise Linear Scaling for Immersive Video Coding

  • Dohyeon Park,
  • Sung-Gyun Lim,
  • Kwan-Jung Oh,
  • Gwangsoon Lee,
  • Jae-Gon Kim

DOI
https://doi.org/10.1109/ACCESS.2022.3140537
Journal volume & issue
Vol. 10
pp. 4483 – 4494

Abstract

Read online

Moving Picture Experts Group (MPEG) is developing a standard for immersive video coding called MPEG Immersive Video (MIV) and is releasing a reference software called Test Model for Immersive Video (TMIV) in the standardization process. The TMIV efficiently compresses an immersive video comprising a set of texture and depth views acquired using multiple cameras within a limited 3D viewing space. Moreover, it affords a rendered view of an arbitrary view position and orientation with six degrees of freedom. However, the existing depth quantization applied to depth atlas in TMIV is insufficient since the reconstructed depth is crucial for achieving the required quality of a rendered viewport. To address this issue, we propose a nonlinear depth quantization method that allocates more codewords to a depth subrange with a higher occurrence of depth values located at edge regions, which are important in terms of the rendered view quality. We implement the proposed nonlinear quantization based on piecewise linear scaling considering the computational complexity and bitstream overhead. The experimental results show that the proposed method yields PSNR-based Bjøntegaard delta rate gains of 5.2% and 4.9% in the end-to-end performance for High- and Low-bitrate ranges, respectively. Moreover, subjective quality improvement is mainly observed at the object boundaries of the rendered viewport. The proposed nonlinear quantization method has been adopted into the TMIV as a candidate standard technology for the next MIV edition.

Keywords