Skip to main content

PMT: Partial-Modality Translation Based on Diffusion Models for Prostate Magnetic Resonance and Ultrasound Image Registration

  • Conference paper
  • First Online:
Medical Image Understanding and Analysis (MIUA 2024)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14860))

Included in the following conference series:

  • 607 Accesses

Abstract

In diagnosing prostate cancer, urologists frequently rely on the Magnetic Resonance Image (MRI) to pinpoint suspicious areas, followed by biopsies guided by ultrasound (US). Aligning MRI and US images is necessary during the procedure, and this task is currently performed manually by highly skilled experts only. Additionally, the notable differences between MR and US images also pose challenges for AI algorithms in the registration process. Therefore, this paper proposes a novel method aimed at bridging the modality gap between MR and US images, thereby easing the registration problem associated with them. We combine the strengths of a 3D Diffusion Model and a Generative Adversarial Network (GAN) to autonomously translate both MR and US images into an intermediate pseudo modality. This marks the initial endeavor to achieve such partial modality translation between 3D MR and US images. Compared to existing state-of-the-art fully modality translation techniques, our method visibly preserves the original image details while producing images with more similar textures. This improvement is objectively evident in our 33.67% reduction in the Fréchet Inception Distance (FID), which is more than double the 15.86% reduction achieved by the existing method. Additionally, our method achieves an impressive 54.90% reduction in the Kernel Inception Distance (KID), surpassing the 17.65% reduction attained by existing methods by more than threefold. Furthermore, we also provide evidence to illustrate that these enhancements significantly improve the effectiveness of the downstream registration task. Additionally, by exclusively employing modality-translated results to derive the warping map, along with conducting the actual warping on the original images, we effectively address the well-known “hallucination issue” in AI-generated medical images.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Al-Karawi, A.S., et al.: Revealing the urinary microbiota in prostate cancer: a comprehensive review unveiling insights into pathogenesis and clinical application. Al-Salam J. Med. Sci. 3(1), 45–54 (2024)

    Article  Google Scholar 

  2. Bińkowski, M., et al.: Demystifying MMD GANs. In: International Conference on Learning Representations (2018)

    Google Scholar 

  3. Bratt, O., et al.: Population-based organised prostate cancer testing: results from the first invitation of 50-year-old men. Eur. Urol. 85(3), 207–214 (2024)

    Article  Google Scholar 

  4. Chen, Y., et al.: MR to ultrasound image registration with segmentation-based learning for HDR prostate brachytherapy. Med. Phys. 48(6), 3074–3083 (2021)

    Article  Google Scholar 

  5. El Mansouri, O., et al.: Fusion of magnetic resonance and ultrasound images for endometriosis detection. IEEE Trans. Image Process. 29, 5324–5335 (2020)

    Article  Google Scholar 

  6. Gaudreau-Balderrama, A.: Multi-modal image registration. Boston University (2012)

    Google Scholar 

  7. Goodfellow, I., et al.: Generative adversarial nets. Adv. Neural Inf. Process. Syst. 27 (2014)

    Google Scholar 

  8. Henry, J., Natalie, T., Madsen, D.: Pix2pix GAN for image-to-image translation, pp. 1–5. Research Gate Publication (2021)

    Google Scholar 

  9. Heusel, M., et al.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. Adv. Neural Inf. Process. Syst. 30 (2017)

    Google Scholar 

  10. Ho, J., et al.: Denoising diffusion probabilistic models. Adv. Neural. Inf. Process. Syst. 33, 6840–6851 (2020)

    Google Scholar 

  11. Jiao, J., et al.: Self-supervised ultrasound to MRI fetal brain image synthesis. IEEE Trans. Med. Imaging 39(12), 4413–4424 (2020)

    Article  Google Scholar 

  12. Kim, B., et al.: Unpaired image-to-image translation via neural schrödinger bridge. In: The Twelfth International Conference on Learning Representations (2023)

    Google Scholar 

  13. Léonard, C.: A survey of the Schrodinger problem and some of its connections with optimal transport. Dyn. Syst. 34(4), 1533–1574 (2014)

    MathSciNet  Google Scholar 

  14. Liu, Y., et al.: CT synthesis from MRI using multi-cycle GAN for head-and-neck radiation therapy. Comput. Med. Imaging Graph. 91, 101953 (2021)

    Article  Google Scholar 

  15. Ma, X., et al.: Unsupervised image fusion using deep image priors. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 2301–2305. IEEE (2022)

    Google Scholar 

  16. Thomson, B.R., et al.: MR-to-US registration using multiclass segmentation of hepatic vasculature with a reduced 3D U-Net. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12263, pp. 275–284. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-59716-0_27

    Chapter  Google Scholar 

  17. Tian, X., et al.: Optimal transport-based graph matching for 3d retinal OCT image registration. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 2791–2795. IEEE (2022)

    Google Scholar 

  18. Tong, A., et al.: Conditional flow matching: simulation-free dynamic optimal transport 2(3). arXiv preprint arXiv:2302.00482 (2023)

  19. Ulyanov, D., Vedaldi, A., Lempitsky, V.: Deep image prior. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9446–9454 (2018)

    Google Scholar 

  20. Ustyuzhaninov, I., et al.: Texture synthesis using shallow convolutional networks with random filters. arXiv preprint arXiv:1606.00021 (2016)

  21. Wang, G., et al.: Deep generative learning via schrödinger bridge. In: International Conference on Machine Learning, pp. 10794–10804. PMLR (2021)

    Google Scholar 

  22. Wang, R., et al.: Bright-field to fluorescence microscopy image translation for cell nuclei health quantification. Biol. Imag. 3, e12 (2023)

    Article  Google Scholar 

  23. Zhu, J.Y., et al.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223–2232 (2017)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xudong Ma .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ma, X., Anantrasirichai, N., Bolomytis, S., Achim, A. (2024). PMT: Partial-Modality Translation Based on Diffusion Models for Prostate Magnetic Resonance and Ultrasound Image Registration. In: Yap, M.H., Kendrick, C., Behera, A., Cootes, T., Zwiggelaar, R. (eds) Medical Image Understanding and Analysis. MIUA 2024. Lecture Notes in Computer Science, vol 14860. Springer, Cham. https://doi.org/10.1007/978-3-031-66958-3_21

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-66958-3_21

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-66957-6

  • Online ISBN: 978-3-031-66958-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics