IEEE Access (Jan 2023)

Enhancing Pedestrian Group Detection and Tracking Through Zone-Based Clustering

  • Mingzuoyang Chen,
  • Shadi Banitaan,
  • Mina Maleki

DOI
https://doi.org/10.1109/ACCESS.2023.3336592
Journal volume & issue
Vol. 11
pp. 132162 – 132179

Abstract

Read online

Advancements in self-driving car technology have the potential to revolutionize transportation by enhancing safety, efficiency, and accessibility. Nonetheless, the successful integration of autonomous vehicles into our urban landscapes necessitates robust and reliable pedestrian detection and tracking systems. As we frequently observe pedestrians moving together, tracking them as a group becomes a beneficial approach, mitigating occlusion and enhancing both the accuracy and speed of object detection and tracking. However, utilizing a human-view camera in an autonomous vehicle presents challenges as pedestrians occupy varied fields of view. In some instances, pedestrians closer to the camera may overlap with those farther away, as seen from the camera’s viewpoint, which causes the mis-groupings to happen. To address these challenges, we proposed a strategy to divide the image into distinct zones and perform grouping within each, significantly minimizing mis-groupings. First, an object detection method extracted pedestrians and their bounding boxes from an image. Second, zone detection was applied to separate the image into several zones. Third, clustering methods were applied to detect pedestrian groups within each zone. Last, the object tracking method was utilized to track pedestrian groups. We repeated the process over a ten-frame sequence to achieve better performance, with object detection executed in the first frame and object tracking in the remaining nine frames. The comparison of processing times of different group detection methods indicated that tracking pedestrian groups is more time-efficient than tracking individuals and achieved a 4.5% to 14.1% improvement. Furthermore, according to the Adjusted Rand Index (ARI) evaluation metric, our proposed zone-based group detection method outperforms the other commonly used approaches by achieving scores of 0.635 on the MOT17 dataset and 0.781 on the KITTI dataset. In addition, the proposed approach surpasses the other approaches in addressing scenarios where individuals from different fields of view intersect with each other.

Keywords