Computational Visual Media (Nov 2023)

A unified multi-view multi-person tracking framework

  • Fan Yang,
  • Shigeyuki Odashima,
  • Sosuke Yamao,
  • Hiroaki Fujimoto,
  • Shoichi Masui,
  • Shan Jiang

DOI
https://doi.org/10.1007/s41095-023-0334-8
Journal volume & issue
Vol. 10, no. 1
pp. 137 – 160

Abstract

Read online

Abstract Despite significant developments in 3D multi-view multi-person (3D MM) tracking, current frameworks separately target footprint tracking, or pose tracking. Frameworks designed for the former cannot be used for the latter, because they directly obtain 3D positions on the ground plane via a homography projection, which is inapplicable to 3D poses above the ground. In contrast, frameworks designed for pose tracking generally isolate multi-view and multi-frame associations and may not be sufficiently robust for footprint tracking, which utilizes fewer key points than pose tracking, weakening multi-view association cues in a single frame. This study presents a unified multi-view multi-person tracking framework to bridge the gap between footprint tracking and pose tracking. Without additional modifications, the framework can adopt monocular 2D bounding boxes and 2D poses as its input to produce robust 3D trajectories for multiple persons. Importantly, multi-frame and multi-view information are jointly employed to improve association and triangulation. Our framework is shown to provide state-of-the-art performance on the Campus and Shelf datasets for 3D pose tracking, with comparable results on the WILDTRACK and MMPTRACK datasets for 3D footprint tracking.

Keywords