Skip to main content
Log in

Probabilistic medical image imputation via deep adversarial learning

  • Original Article
  • Published:
Engineering with Computers Aims and scope Submit manuscript

Abstract

The ability to impute missing images from a sequence of medical images plays an important role in enabling the detection, diagnosis and treatment of disease. Motivated by this, in this manuscript we propose a novel, probabilistic deep-learning algorithm for imputing images. Within this approach, given a sequence of contrast enhanced CT images, we train a generative adversarial network (GAN) to learn the underlying probabilistic relation between these images. Thereafter, given all but one member from a sequence, we infer the probability distribution of the missing image using Bayesian inference. We make the inference problem computationally tractable by mapping it to the low-dimensional latent space of the GAN. Thereafter, we use Markov Chain Monte Carlo (MCMC) techniques to learn and sample the inferred distribution. Moreover, we propose a novel style loss unique to contrast-enhanced computed tomography (CECT) imaging to improve the texture of the generated images, and apply these techniques to infer missing CECT images of renal masses collected during an IRB-approved retrospective study. In doing so, we demonstrate how the ability to infer the probability distribution of the missing image, as opposed to a single image recovery, can be used by the end-user to quantify the reliability of the imputed results.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Explore related subjects

Discover the latest articles and news from researchers in related subjects, suggested using machine learning.

Notes

  1. https://github.com/RaghebRaad400/Probabilistic-Medical-Image-Imputation-via-Deep-Adversarial-Learning

References

  1. Oglevee C, Pianykh O (2015) Losing images in digital radiology: more than you think. J Digit Imaging 28(3):264–271

    Article  Google Scholar 

  2. Dalca AV, Bouman KL, Freeman WT, Rost NS, Sabuncu MR, Golland P (2018) Medical image imputation from image collections. IEEE Trans Med Imaging 38(2):504–514

    Article  Google Scholar 

  3. Xia Y, Zhang L, Ravikumar N, Attar R, Piechnik SK, Neubauer S, Petersen SE, Frangi AF (2021) Recovering from missing data in population imaging-cardiac mr image imputation via conditional generative adversarial nets. Med Image Anal 67:101812

    Article  Google Scholar 

  4. Heilbrun ME, Remer EM, Casalino DD, Beland MD, Bishoff JT, Blaufox MD, Coursey CA, Goldfarb S, Harvin HJ, Nikolaidis P (2015) Acr appropriateness criteria indeterminate renal mass. J Am Coll Radiol 12(4):333–341

    Article  Google Scholar 

  5. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. Adv Neural Inf Process Syst 27:3

    Google Scholar 

  6. Gamerman D, Lopes HF (2006) Markov chain Monte Carlo: stochastic simulation for bayesian inference. CRC Press, Hoboken

    Book  MATH  Google Scholar 

  7. Zhang L, Pereañez M, Bowles C, Piechnik S, Neubauer S, Petersen S, Frangi A (2019) Missing slice imputation in population CMR imaging via conditional generative adversarial nets. In: Shen D, Liu T, Peters TM, Staib LH, Essert C, Zhou S, Yap P-T, Khan A (eds) Medical image computing and computer assisted intervention—MICCAI 2019. Springer, Cham, pp 651–659

    Chapter  Google Scholar 

  8. Dinov ID, Herting MM, Chen G-Z, Kim H, Toga AW, Sepehrband F (2020) Imputation strategy for reliable regional mri morphological measurements. Neuroinformatics 18(1):59–70. https://doi.org/10.1007/S12021-019-09426-X

    Article  Google Scholar 

  9. Zhu J-Y, Park T, Isola P, Efro AA (2017) Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE international conference on computer vision, pp 2223–2232

  10. Choi Y, Choi M, Kim M, Ha J-W, Kim S, Choo J (2018) Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In: Proceedings of the ieee conference on computer vision and pattern recognition, pp 8789–8797

  11. Lee D, Kim J, Moon W-J, Ye JC (2019) Collagan: collaborative gan for missing image data imputation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2487–2496

  12. Patel DV, Oberai AA (2020) GAN-based priors for quantifying uncertainty. https://doi.org/10.13140/RG.2.2.28806.32322. arXiv:2003.12597.

  13. Patel D, Oberai AA (2019) Bayesian inference with generative adversarial network priors. arXiv:1907.09987

  14. Almahairi A, Rajeshwar S, Sordoni A, Bachman P, Courville A (2018) Augmented cyclegan: learning many-to-many mappings from unpaired data. In: International conference on machine learning, pp 195–204 , PMLR

  15. Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: International conference on machine learning, pp 214–223, PMLR

  16. Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville A (2017) Improved training of wasserstein gans. arXiv:1704.00028

  17. Dashti M, Stuart AM (2016) The bayesian approach to inverse problems. Handb Uncertain Quantif 2016:1–118

    Google Scholar 

  18. Gatys L, Ecker AS, Bethge M (2015) Texture synthesis using convolutional neural networks. Adv Neural Inf Process Syst 28:262–270

    Google Scholar 

  19. Xian W, Sangkloy P, Agrawal V, Raj A, Lu J, Fang C, Yu F, Hays J (2018) Texturegan: Controlling deep image synthesis with texture patches. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8456–8465

  20. Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2414–2423

  21. Karras T, Laine S, Aittala M, Hellsten J, Lehtinen J, Aila T (2020) Analyzing and improving the image quality of stylegan. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8110–8119

  22. Oberai A, Varghese B, Cen S, Angelini T, Hwang D, Gill I, Aron M, Lau C, Duddalwar V (2020) Deep learning based classification of solid lipid-poor contrast enhancing renal masses using contrast enhanced ct. Br J Radiol 93(1111):20200002

    Article  Google Scholar 

  23. Yap FY, Varghese BA, Cen SY, Hwang DH, Lei X, Desai B, Lau C, Yang LL, Fullenkamp AJ, Hajian S (2021) Shape and texture-based radiomics signature on ct effectively discriminates benign from malignant renal masses. Eur Radiol 31(2):1011–1021

    Article  Google Scholar 

  24. Zwanenburg A, Leger S, Vallières M, Löck S (2016) Image biomarker standardisation initiative. arXiv:1612.07003

  25. Betancourt M (2017) A Conceptual introduction to Hamiltonian Monte Carlo. arXiv:1701.02434

  26. Mattingly JC, Pillai NS, Stuart AM (2012) Diffusion limits of the random walk Metropolis algorithm in high dimensions. Ann Appl Probab 22(3):881–930

    Article  MathSciNet  MATH  Google Scholar 

  27. Roy V (2020) Convergence diagnostics for markov chain monte carlo. Annu Rev Stat Appl 7(1):387–412

    Article  MathSciNet  Google Scholar 

  28. Gelman A, Rubin DB (1992) Inference from iterative simulation using multiple sequences. Stat Sci 7(4):457–472. https://doi.org/10.1214/ss/1177011136

    Article  MATH  Google Scholar 

  29. Hoffman MD, Gelman A (2014) The No-U-turn sampler: adaptively setting path lengths in Hamiltonian Monte Carlo. Technical report. http://mcmc-jags.sourceforge.net

  30. Dillon JV, Langmore I, Tran D, Brevdo E, Vasudevan S, Moore D, Patton B, Alemi A, Hoffman M, Saurous RA (2017) TensorFlow distributions. arXiv:1711.10604

  31. Andrieu C, Thoms J (2008) A tutorial on adaptive MCMC. Stat Comput 18(4):343–373. https://doi.org/10.1007/s11222-008-9110-y

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

The support from ARO grant W911NF2010050 and the Ming-Hsieh Institute at USC is acknowledged.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Assad A. Oberai.

Ethics declarations

Conflict of interest

The authors have no relevant financial or non-financial interests to disclose.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendices

Appendix A Convergence of MCMC algorithm

MCMC algorithms are known to suffer from challenges when used in high-dimensional spaces, where the mass of the target density is typically concentrated in narrow regions on a lower dimensional manifold [25]. To fully explore the regions of interest, the requirements on the length of the Markov chains [26] can make the algorithm computationally infeasible. Thus, posing the inference problem on the lower-dimensional latent space can alleviate this issue.

A number of diagnostic tools are available to analyse the convergence of the MCMC algorithm, and thus determine the termination length of the generated Markov chains. We direct interested readers to [27] for a summary of such techniques. In the present work, we use the Gelman-Rubin diagnostic [28] which estimates the convergence by considering multiple Markov chains and evaluating the between-chains and within-chains variances.

We consider M chains of length N, each of which is generated by the MCMC algorithm from different random initial points. Let \(\mu _m\) and \(\sigma _m^2\) denote the sample mean and variance of the mth chain, and \(\mu\) denote the overall mean across all chains, i.e., \(\mu = \sum _{m=1}^M \mu _m/M\). Then, we estimate the within-chain variance W, the between-chain variance B and the pooled variance \(\widehat{V}\) as

$$\begin{aligned} \begin{aligned} W&= \frac{1}{M} \sum _{m=1}^M \sigma _m^2, \quad B = \frac{N}{M-1}\sum _{m=1}^M(\mu _m - \mu )^2,\\ \widehat{V}&= \frac{N-1}{N} W + \frac{B}{N}. \end{aligned} \end{aligned}$$
(A1)

Finally, we evaluate the potential scale reduction factor \(\widehat{R} = \sqrt{\widehat{V}/W}\). Assuming that the initial points of the chains were sampled from an over-dispersed distribution compared to the target distribution, \(\widehat{V}\) is expected to overestimate the variance of the target distribution, while W underestimates it. Thus, the closer that value of \(\widehat{R}\) is to 1, the more assured we are about the convergence of the chains.

To demonstrate the utility of this tool, we consider the chains generated to impute the missing images at the 4 time-points for Subject 1. At each time-point, we use \(M=4\) chains for each of the lengths N = 256 K, 512 K, 1024 K. Since the chains are generated for latent variable \(z \in {\mathbb {R}}^{N_z}\), we obtain a vector \(\widehat{R} \in {\mathbb {R}}^{N_z}\) for each configuration. To simplify the analysis, we condense this vector to a scalar by considering the dimensional mean ± standard deviation of \(\widehat{R}\), which is listed in Table 1. Note that these scalar value moves closer to 1 as N increases, indicating convergence. In practice, a value of \(\widehat{R} < 1.2\) is considered as a good termination threshold. To balance the convergence of the chains and the associated computational cost, we use \(N=1024\) K for all results presented in this work.

Table 1 \(\widehat{R}\) values of Markov chains in z used to impute the images at the four time-points for Subject 1

Appendix B Architecture and hyper-parameters

Table 2 Generator architecture
Table 3 Discriminator architecture

We use the axial slice of the 3000 CECT images for training. The WGAN-GP models, whose architectures are described in Tables 2 and 3, are trained with the gradient penalty parameter set to 10. We use the ADAM optimizer with learning rate of \(2 \times 10^{-4}\) and momentum parameters \(\beta _1 = 0.0, \beta _2 = 0.9\). We perform 5 gradient updates of critic per gradient update of generator and train both networks in TensorFlow with a batch size of 64. For posterior inference we sample Markov chain using Hamiltonian Monte Carlo (HMC) with No-U Turn Sampler (NUTS) [29] and implement it in TensorFlow Probability [30]. We use initial step size of 1.0 for HMC and adapt it following [31] based on the target acceptance probability. A burn-in period of 50% is used for all HMC simulations. These hyper-parameters are chosen to ensure the convergence of the chains.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Raad, R., Patel, D., Hsu, CC. et al. Probabilistic medical image imputation via deep adversarial learning. Engineering with Computers 38, 3975–3986 (2022). https://doi.org/10.1007/s00366-022-01712-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00366-022-01712-8

Keywords