Applied Sciences (Mar 2024)

Sparsity-Robust Feature Fusion for Vulnerable Road-User Detection with 4D Radar

  • Leon Ruddat,
  • Laurenz Reichardt,
  • Nikolas Ebert,
  • Oliver Wasenmüller

DOI
https://doi.org/10.3390/app14072781
Journal volume & issue
Vol. 14, no. 7
p. 2781

Abstract

Read online

Detecting vulnerable road users is a major challenge for autonomous vehicles due to their small size. Various sensor modalities have been investigated, including mono or stereo cameras and 3D LiDAR sensors, which are limited by environmental conditions and hardware costs. Radar sensors are a low-cost and robust option, with high-resolution 4D radar sensors being suitable for advanced detection tasks. However, they involve challenges such as few and irregularly distributed measurement points and disturbing artifacts. Learning-based approaches utilizing pillar-based networks show potential in overcoming these challenges. However, the severe sparsity of radar data makes detecting small objects with only a few points difficult. We extend a pillar network with our novel Sparsity-Robust Feature Fusion (SRFF) neck, which combines high- and low-level multi-resolution features through a lightweight attention mechanism. While low-level features aid in better localization, high-level features allow for better classification. As sparse input data are propagated through a network, the increasing effective receptive field leads to feature maps of different sparsities. The combination of features with different sparsities improves the robustness of the network for classes with few points.

Keywords