Abstract
Intra-operative non-enhanced computed tomography (CT) has grown as a prominent tool in interventional surgeries for liver tumor by offering a real-time, more precise imaging guidance over its pre-operative enhanced counterpart as well as intra-operative ultrasound. Though deep learning has demonstrated promising performance in various medical image tasks, the challenges for liver tumor segmentation in the context of intra-operative non-enhanced CT remain due to the scarcity of accurate pixel-level-labeled tumor data and low visibility of tumor without contrast enhancement. In this paper, based on a pre-assumption of consistent relative spatial relation between liver and liver tumor at pre- and intra-surgical stages, we propose a two-step (segmentation - registration) framework to propagate liver tumor segmentation mask from pre-operative enhanced CT to corresponding intra-operative non-enhanced CT via nonrigid registration. A Multi-scale Bridged spatial attention block embedded in the popular “U-shaped” network architecture termed MsBs-Unet is introduced to better accommodate variations of tumor during segmentation. We show that MsBs-Unet uniformly achieves superior performances against other baselines on liver and liver tumor segmentation task in terms of Dice score and HD95 metrics on our in-house data containing a cohort of intra-operative non-enhanced CT scans and Medical Segmentation Decathlon task03 liver tumor dataset respectively. Final registration results were evaluated via tumor landmarks matching on 20 in-house intra-subject paired CT scans for both intensity and feature-based methods. Specifically for the latter, mean target registration error is 3.3 mm, and the overall intra-operative process takes around 48 s per case. The resulting tumor segmentation show gratifying promise of applying our framework for non-enhanced CT-based liver intervention.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bozinov, O., Burkhardt, J.K., Fischer, C.M., Kockro, R.A., Bernays, R.L., Bertalanffy, H.: Advantages and limitations of intraoperative 3D ultrasound in neurosurgery. Technical note. In: Pamir, M., Seifert, V., Kiris, T. (eds.) Intraoperative Imaging. Acta Neurochirurgica Supplementum, vol. 109. Springer, Vienna (2011). https://doi.org/10.1007/978-3-211-99651-5_30 (2011)
Celaya, A., Diaz, A., Balsells, A., Riviere, B., Fuentes, D.: A weighted normalized boundary loss for reducing the Hausdorff distance in medical imaging segmentation. arXiv preprint arXiv:2302.03868 (2023)
Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)
Fang, H.Y., et al.: Efficacy and safety of preoperative vs intraoperative computed tomography-guided lung tumor localization: a randomized controlled trial. Front. Surg. 8, 809908 (2022)
Hatamizadeh, A., Nath, V., Tang, Y., Yang, D., Roth, H.R., Xu, D.: Swin UNETR: Swin transformers for semantic segmentation of brain tumors in MRI images. In: Crimi, A., Bakas, S. (eds.) Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries. BrainLes 2021. LNCS, vol. 12962. Springer, Cham (2022). https://doi.org/10.1007/978-3-031-08999-2_22
Hatamizadeh, A., et al.: UNETR: transformers for 3d medical image segmentation. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 574–584 (2022)
Huang, Z., et al.: ISA-Net: improved spatial attention network for PET-CT tumor segmentation. Comput. Methods Programs Biomed. 226, 107129 (2022)
Ibtehaz, N., Rahman, M.S.: MultiResUNet: rethinking the u-net architecture for multimodal biomedical image segmentation. Neural Netw. 121, 74–87 (2020)
Isensee, F., Jaeger, P.F., Kohl, S.A., Petersen, J., Maier-Hein, K.H.: nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 18(2), 203–211 (2021)
Museyko, O., Eisa, F., Hess, A., Schett, G., Kalender, W.A., Engelke, K.: Binary segmentation masks can improve intrasubject registration accuracy of bone structures in CT images. Ann. Biomed. Eng. 38, 2464–2472 (2010)
Nitsch, J., et al.: Automatic and efficient MRI-US segmentations for improving intraoperative image fusion in image-guided neurosurgery. NeuroImage: Clin. 22, 101766 (2019)
Oguro, S., et al.: MRI signal intensity based B-spline nonrigid registration for pre-and intraoperative imaging during prostate brachytherapy. J. Magn. Reson. Imag. 30(5), 1052–1058 (2009)
Oktay, O., et al.: Attention U-Net: learning where to look for the pancreas. arXiv preprint arXiv:1804.03999 (2018)
Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-24574-4_28
Roy, A.G., Navab, N., Wachinger, C.: Concurrent spatial and channel ‘squeeze & excitation’ in fully convolutional networks. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11070, pp. 421–429. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00928-1_48
Simpson, A.L., et al.: A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv preprint arXiv:1902.09063 (2019)
Sung, H., et al.: Global cancer statistics 2020: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: Cancer J. Clin. 71(3), 209–249 (2021)
Zhao, Y., Chen, J., Zhang, Z., Zhang, R.: BA-Net: bridge attention for deep convolutional neural networks. In: Avidan, S., Brostow, G., Cissé, M., Farinella, G.M., Hassner, T. (eds.) Computer Vision – ECCV 2022. ECCV 2022. LNCS, vol. 13681. Springer, Cham (2022). https://doi.org/10.1007/978-3-031-19803-8_18
Zhou, Z., Siddiquee, M.M.R., Tajbakhsh, N., Liang, J.: UNet++: redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 39(6), 1856–1867 (2019)
Acknowledgment
The study was supported by National Natural Science Foundation of China (81827805, 82130060, 61821002, 92148205), National Key Research and Development Program (2018YFA0704100, 2018YFA0704104). The project was funded by China Postdoctoral Science Foundation (2021M700772), Zhuhai Industry-University-Research Collaboration Program (ZH22017002210011PWC), Jiangsu Provincial Medical Innovation Center (CXZX202219), Collaborative Innovation Center of Radiation Medicine of Jiangsu Higher Education Institutions, and Nanjing Life Health Science and Technology Project (202205045). The funding sources had no role in the writing of the report, or decision to submit the paper for publication.
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Lyu, P. et al. (2023). Registration-Propagated Liver Tumor Segmentation for Non-enhanced CT-Based Interventions. In: Lu, H., et al. Image and Graphics . ICIG 2023. Lecture Notes in Computer Science, vol 14358. Springer, Cham. https://doi.org/10.1007/978-3-031-46314-3_10
Download citation
DOI: https://doi.org/10.1007/978-3-031-46314-3_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-46313-6
Online ISBN: 978-3-031-46314-3
eBook Packages: Computer ScienceComputer Science (R0)