Abstract
Stereo-tactic neurosurgery planning is a time-consuming and complex task that requires detailed understanding of the patient anatomy and the affected regions in the brain to precisely deliver the treatment and to avoid proximity to any known risk structures. Traditional user interfaces for neurosurgery planning use keyboard and mouse for interaction and visualize the medical data on a screen. Previous research, however, has shown that 3D user interfaces are more intuitive for navigating volumetric data and enable users to understand spatial relations more quickly. Furthermore, new imaging modalities and automated segmentation of relevant structures provide important information to medical experts. However, displaying such information requires frequent context switches or occludes otherwise important information.
In collaboration with medical experts, we analyzed the planning workflow for stereo-tactic neurosurgery interventions and identified two tasks in the process that can be improved: volume exploration and trajectory refinement. In this paper, we present a novel 3D user interface for neurosurgery planning that is implemented using a head-mounted display and a haptic device. The proposed system improves volume exploration with bi-manual interaction to control oblique slicing of volumetric data and reduces visual clutter with the help of haptic guides that enable users to precisely target regions of interest and to avoid proximity to known risk structures.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Abhari, K., Baxter, J.S.H., Chen, E.S., Khan, A.R., Wedlake, C., Peters, T., Eagleson, R., de Ribaupierre, S.: The role of augmented reality in training the planning of brain tumor resection. In: Liao, H., Linte, C.A., Masamune, K., Peters, T.M., Zheng, G. (eds.) MIAR 2013 and AE-CAI 2013. LNCS, vol. 8090, pp. 241–248. Springer, Heidelberg (2013)
Bériault, S., Al Subaie, F., Collins, D.L., Sadikot, A.F., Pike, G.B.: A multi-modal approach to computer-assisted deep brain stimulation trajectory planning. Int. J. Comput. Assist. Radiol. Surg. 7(5), 687–704 (2012)
Bériault, S., Subaie, F.A., Mok, K., Sadikot, A.F., Pike, G.B.: Automatic trajectory planning of DBS neurosurgery from multi-modal MRI datasets. In: Fichtinger, G., Martel, A., Peters, T. (eds.) MICCAI 2011, Part I. LNCS, vol. 6891, pp. 259–266. Springer, Heidelberg (2011)
Albis, T.D., Haegelen, C., Essert, C., Fernandez-Vidal, S., Lalys, F., Jannin, P.: PyDBS: an automated image processing workflow for deep brain stimulation surgery. Int. J. Comput. Assist. Radiol. Surg. 10(2), 117–128 (2015)
De Momi, E., Caborni, C., Cardinale, F., Castana, L., Casaceli, G., Cossu, M., Antiga, L., Ferrigno, G.: Automatic trajectory planner for stereo-electro-encephalo-graphy procedures: a retrospective study. IEEE Trans. Biomed. Eng. 60(4), 986–993 (2013)
Dormont, D., Seidenwurm, D., Galanaud, D., Cornu, P., Yelnik, J., Bardinet, E.: Neuroimaging and deep brain stimulation. Am. J. Neuroradiol. 31(1), 15–23 (2010)
Eagleson, R., Wucherer, P., Stefan, P., Duschko, Y., de Ribaupierre, S., Vollmar, C., Fallavollita, P., Navab, N.: Collaborative table-top VR display for neurosurgical planning. In: Proceedings of the IEEE Virtual Reality Conference, pp. 169–170 (2015)
Eck, U., Pankratz, F., Sandor, C., Klinker, G., Laga, H.: Precise haptic device co-location for visuo-haptic augmented reality. IEEE Trans. Comput. Graph. Vis. 21(12), 1427–1441 (2015)
Goble, J.C., Hinckley, K., Pausch, R., Snell, J.W., Kassell, N.F.: Two-handed spatial interface tools for neurosurgical planning. IEEE Comput. 1(7), 20–26 (1995)
Huber, M., Pustka, D., Keitler, P., Echtler, F., Klinker, G.: A system architecture for ubiquitous tracking environments. In: Proceedings of the 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, pp. 211–214. IEEE Computer Society, Nara (2007)
Kane, T.R., Levinson, D.A.: Dynamics, Theory and Applications. McGraw-Hill Book Company, New York (1985)
Khlebnikov, R., Kainz, B., Muehl, J., Schmalstieg, D.: Crepuscular rays for tumor accessibility planning. IEEE Trans. Vis. Comput. Graph. 17(12), 2163–2172 (2011)
Preim, B., Bartz, D.: Visualization in Medicine. Morgan Kaufmann, Burlington (2007)
Shamir, R.R., Horn, M., Blum, T., Mehrkens, J.H., Shoshan, Y., Joskowicz, L., Navab, N.: Trajectory planning with augmented reality for improved risk assessment in image-guided keyhole neurosurgery. In: Proceedings of the 8th IEEE International Symposium on Biomedical Imaging, pp. 1873–1876 (2011)
Shamir, R.R., Joskowicz, L., Tamir, I., Dabool, E., Pertman, L., Ben-Ami, A., Shoshan, Y.: Reduced risk trajectory planning in image-guided keyhole neurosurgery. Med. Phys. 39(5), 2885–2895 (2012)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this paper
Cite this paper
Eck, U., Stefan, P., Laga, H., Sandor, C., Fallavollita, P., Navab, N. (2016). Exploring Visuo-Haptic Augmented Reality User Interfaces for Stereo-Tactic Neurosurgery Planning. In: Zheng, G., Liao, H., Jannin, P., Cattin, P., Lee, SL. (eds) Medical Imaging and Augmented Reality. MIAR 2016. Lecture Notes in Computer Science(), vol 9805. Springer, Cham. https://doi.org/10.1007/978-3-319-43775-0_19
Download citation
DOI: https://doi.org/10.1007/978-3-319-43775-0_19
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-43774-3
Online ISBN: 978-3-319-43775-0
eBook Packages: Computer ScienceComputer Science (R0)