PLoS ONE (Jan 2021)

An integrated augmented reality surgical navigation platform using multi-modality imaging for guidance.

  • Harley H L Chan,
  • Stephan K Haerle,
  • Michael J Daly,
  • Jinzi Zheng,
  • Lauren Philp,
  • Marco Ferrari,
  • Catriona M Douglas,
  • Jonathan C Irish

DOI
https://doi.org/10.1371/journal.pone.0250558
Journal volume & issue
Vol. 16, no. 4
p. e0250558

Abstract

Read online

An integrated augmented reality (AR) surgical navigation system that potentially improves intra-operative visualization of concealed anatomical structures. Integration of real-time tracking technology with a laser pico-projector allows the surgical surface to be augmented by projecting virtual images of lesions and critical structures created by multimodality imaging. We aim to quantitatively and qualitatively evaluate the performance of a prototype interactive AR surgical navigation system through a series of pre-clinical studies. Four pre-clinical animal studies using xenograft mouse models were conducted to investigate system performance. A combination of CT, PET, SPECT, and MRI images were used to augment the mouse body during image-guided procedures to assess feasibility. A phantom with machined features was employed to quantitatively estimate the system accuracy. All the image-guided procedures were successfully performed. The tracked pico-projector correctly and reliably depicted virtual images on the animal body, highlighting the location of tumour and anatomical structures. The phantom study demonstrates the system was accurate to 0.55 ± 0.33mm. This paper presents a prototype real-time tracking AR surgical navigation system that improves visualization of underlying critical structures by overlaying virtual images onto the surgical site. This proof-of-concept pre-clinical study demonstrated both the clinical applicability and high precision of the system which was noted to be accurate to <1mm.