IET Computer Vision (Jun 2021)

Object tracking using temporally matching filters

  • Brendan Robeson,
  • Mohammadreza Javanmardi,
  • Xiaojun Qi

DOI
https://doi.org/10.1049/cvi2.12040
Journal volume & issue
Vol. 15, no. 4
pp. 245 – 257

Abstract

Read online

Abstract One of the primary challenges of visual tracking is the variable appearance of the target object. As tracking proceeds, the target object can change its appearance due to illumination changes, rotations, deformations etc. Modern trackers incorporate online updating to learn how the target changes over time. However, they do not use the history of target appearance. To address this shortcoming, we uniquely use domain adaptation with the target appearance history to efficiently learn a temporally matching filter (TMF) during online updating. This TMF emphasizes the persistent features found in different appearances of the target object. It also improves the classification accuracy of the convolutional neural network by assisting the training of the classification layers without incurring the runtime overhead of updating the convolutional layers. Extensive experimental results demonstrate that the proposed TMF‐based tracker, which incorporates domain adaptation with the target appearance history, improves tracking performance on three benchmark video databases (OTB‐50, OTB‐100 and VOT2016) over other online learning trackers. Specifically, it improves the overlap success of VITAL and MDNet by 0.44 % and 1.03 % on the OTB‐100 dataset and improves the accuracy of VITAL and MDNet by 0.55 % and 0.06 % on the VOT2016 dataset, respectively.