Abstract
In diagnosing prostate cancer, urologists frequently rely on the Magnetic Resonance Image (MRI) to pinpoint suspicious areas, followed by biopsies guided by ultrasound (US). Aligning MRI and US images is necessary during the procedure, and this task is currently performed manually by highly skilled experts only. Additionally, the notable differences between MR and US images also pose challenges for AI algorithms in the registration process. Therefore, this paper proposes a novel method aimed at bridging the modality gap between MR and US images, thereby easing the registration problem associated with them. We combine the strengths of a 3D Diffusion Model and a Generative Adversarial Network (GAN) to autonomously translate both MR and US images into an intermediate pseudo modality. This marks the initial endeavor to achieve such partial modality translation between 3D MR and US images. Compared to existing state-of-the-art fully modality translation techniques, our method visibly preserves the original image details while producing images with more similar textures. This improvement is objectively evident in our 33.67% reduction in the Fréchet Inception Distance (FID), which is more than double the 15.86% reduction achieved by the existing method. Additionally, our method achieves an impressive 54.90% reduction in the Kernel Inception Distance (KID), surpassing the 17.65% reduction attained by existing methods by more than threefold. Furthermore, we also provide evidence to illustrate that these enhancements significantly improve the effectiveness of the downstream registration task. Additionally, by exclusively employing modality-translated results to derive the warping map, along with conducting the actual warping on the original images, we effectively address the well-known “hallucination issue” in AI-generated medical images.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Al-Karawi, A.S., et al.: Revealing the urinary microbiota in prostate cancer: a comprehensive review unveiling insights into pathogenesis and clinical application. Al-Salam J. Med. Sci. 3(1), 45–54 (2024)
Bińkowski, M., et al.: Demystifying MMD GANs. In: International Conference on Learning Representations (2018)
Bratt, O., et al.: Population-based organised prostate cancer testing: results from the first invitation of 50-year-old men. Eur. Urol. 85(3), 207–214 (2024)
Chen, Y., et al.: MR to ultrasound image registration with segmentation-based learning for HDR prostate brachytherapy. Med. Phys. 48(6), 3074–3083 (2021)
El Mansouri, O., et al.: Fusion of magnetic resonance and ultrasound images for endometriosis detection. IEEE Trans. Image Process. 29, 5324–5335 (2020)
Gaudreau-Balderrama, A.: Multi-modal image registration. Boston University (2012)
Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27 (2014)
Henry, J., Natalie, T., Madsen, D.: Pix2pix GAN for image-to-image translation, pp. 1–5. Research Gate Publication (2021)
Heusel, M., et al.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process. Syst. 30 (2017)
Ho, J., et al.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840–6851 (2020)
Jiao, J., et al.: Self-supervised ultrasound to MRI fetal brain image synthesis. IEEE Trans. Med. Imaging 39(12), 4413–4424 (2020)
Kim, B., et al.: Unpaired image-to-image translation via neural schrödinger bridge. In: The Twelfth International Conference on Learning Representations (2023)
Léonard, C.: A survey of the Schrodinger problem and some of its connections with optimal transport. Dyn. Syst. 34(4), 1533–1574 (2014)
Liu, Y., et al.: CT synthesis from MRI using multi-cycle GAN for head-and-neck radiation therapy. Comput. Med. Imaging Graph. 91, 101953 (2021)
Ma, X., et al.: Unsupervised image fusion using deep image priors. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 2301–2305. IEEE (2022)
Thomson, B.R., et al.: MR-to-US registration using multiclass segmentation of hepatic vasculature with a reduced 3D U-Net. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12263, pp. 275–284. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-59716-0_27
Tian, X., et al.: Optimal transport-based graph matching for 3d retinal OCT image registration. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 2791–2795. IEEE (2022)
Tong, A., et al.: Conditional flow matching: simulation-free dynamic optimal transport 2(3). arXiv preprint arXiv:2302.00482 (2023)
Ulyanov, D., Vedaldi, A., Lempitsky, V.: Deep image prior. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9446–9454 (2018)
Ustyuzhaninov, I., et al.: Texture synthesis using shallow convolutional networks with random filters. arXiv preprint arXiv:1606.00021 (2016)
Wang, G., et al.: Deep generative learning via schrödinger bridge. In: International Conference on Machine Learning, pp. 10794–10804. PMLR (2021)
Wang, R., et al.: Bright-field to fluorescence microscopy image translation for cell nuclei health quantification. Biol. Imag. 3, e12 (2023)
Zhu, J.Y., et al.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223–2232 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Ma, X., Anantrasirichai, N., Bolomytis, S., Achim, A. (2024). PMT: Partial-Modality Translation Based on Diffusion Models for Prostate Magnetic Resonance and Ultrasound Image Registration. In: Yap, M.H., Kendrick, C., Behera, A., Cootes, T., Zwiggelaar, R. (eds) Medical Image Understanding and Analysis. MIUA 2024. Lecture Notes in Computer Science, vol 14860. Springer, Cham. https://doi.org/10.1007/978-3-031-66958-3_21
Download citation
DOI: https://doi.org/10.1007/978-3-031-66958-3_21
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-66957-6
Online ISBN: 978-3-031-66958-3
eBook Packages: Computer ScienceComputer Science (R0)