International Journal of Advanced Robotic Systems (Nov 2008)

A Monocular Pointing Pose Estimator for Gestural Instruction of a Mobile Robot

  • Andrea Scheidig,
  • Christian Martin,
  • Steffen Mueler,
  • Horst-Michael Gross,
  • Jan Richarz

Journal volume & issue
Vol. 4, no. 1

Abstract

Read online

We present an important aspect of our human-robot communication interface which is being developed in the context of our long-term research framework PERSES dealing with highly interactive mobile companion robots. Based on a multi-modal people detection and tracking system, we present a hierarchical neural architec- ture that estimates a target point at the floor indicated by a pointing pose, thus enabling a user to navigate a mo- bile robot to a specific target position in his local surroundings by means of pointing. In this context, we were especially interested in determining whether it is possible to accomplish such a target point estimator using only monocular images of low-cost cameras. The estimator has been implemented and experimentally investigated on our mobile robotic assistant HOROS. Although only monocular image data of relatively poor quality were util- ized, the estimator accomplishes a good estimation performance, achieving an accuracy better than that of a hu- man viewer on the same data. The achieved recognition results demonstrate that it is in fact possible to realize a user-independent pointing direction estimation using monocular images only, but further efforts are necessary to improve the robustness of this approach for everyday application.