J Neurol Surg B Skull Base 2021; 82(S 03): e268-e270
DOI: 10.1055/s-0040-1701675
Original Article

Integrating Stereoscopic Video with Modular 3D Anatomic Models for Lateral Skull Base Training

1   Department of Otolaryngology-Head and Neck Surgery, University of Arizona College of Medicine, Tucson, Arizona, United States
,
Saurabh Jain
2   Department of Systems and Industrial Engineering, University of Arizona, Tucson, Arizona, United States
,
Young-Jun Son
2   Department of Systems and Industrial Engineering, University of Arizona, Tucson, Arizona, United States
,
Kaith Almefty
3   Division of Neurotology and Lateral Skull Base Surgery, Barrow Brain and Spine, Barrow Neurological Institute, Phoenix, Arizona, United States
,
Michael T. Lawton
3   Division of Neurotology and Lateral Skull Base Surgery, Barrow Brain and Spine, Barrow Neurological Institute, Phoenix, Arizona, United States
,
Shawn M. Stevens
3   Division of Neurotology and Lateral Skull Base Surgery, Barrow Brain and Spine, Barrow Neurological Institute, Phoenix, Arizona, United States
4   The Arizona Ear Institute, Phoenix, Arizona, United States
› Author Affiliations
Preview

Abstract

Introduction Current virtual reality (VR) technology allows the creation of instructional video formats that incorporate three-dimensional (3D) stereoscopic footage.Combined with 3D anatomic models, any surgical procedure or pathology could be represented virtually to supplement learning or surgical preoperative planning. We propose a standalone VR app that allows trainees to interact with modular 3D anatomic models corresponding to stereoscopic surgical videos.

Methods Stereoscopic video was recorded using an OPMI Pentero 900 microscope (Zeiss, Oberkochen, Germany). Digital Imaging and Communications in Medicine (DICOM) images segmented axial temporal bone computed tomography and each anatomic structure was exported separately. 3D models included semicircular canals, facial nerve, sigmoid sinus and jugular bulb, carotid artery, tegmen, canals within the temporal bone, cochlear and vestibular aqueducts, endolymphatic sac, and all branches for cranial nerves VII and VIII. Finished files were imported into the Unreal Engine. The resultant application was viewed using an Oculus Go.

Results A VR environment facilitated viewing of stereoscopic video and interactive model manipulation using the VR controller. Interactive models allowed users to toggle transparency, enable highlighted segmentation, and activate labels for each anatomic structure. Based on 20 variable components, a value of 1.1 × 1012 combinations of structures per DICOM series was possible for representing patient-specific anatomy in 3D.

Conclusion This investigation provides proof of concept that a hybrid of stereoscopic video and VR simulation is possible, and that this tool may significantly aid lateral skull base trainees as they learn to navigate a complex 3D surgical environment. Future studies will validate methodology.



Publication History

Received: 26 September 2019

Accepted: 24 December 2019

Article published online:
13 March 2020

© 2020. Thieme. All rights reserved.

Georg Thieme Verlag KG
Rüdigerstraße 14, 70469 Stuttgart, Germany