Generating future fundus images for early age-related macular degeneration based on generative adversarial networks

https://doi.org/10.1016/j.cmpb.2022.106648Get rights and content

Highlights

  • The first attempt to generate future fundus images based on current fundus images for early age-related macular degeneration (AMD) patients based on deep learning model.

  • Exploit the drusen segmentation mask for improving the performance.

  • Introduce a GAN-based model with two discriminators for preserving the identity and utilizing drusen masks.

  • Develop a fundus dataset for our task.

Abstract

Background and objective: Age-related macular degeneration (AMD) is one of the most common diseases that can lead to blindness worldwide. Recently, various fundus image analyzing studies are done using deep learning methods to classify fundus images to aid diagnosis and monitor AMD disease progression. But until now, to the best of our knowledge, no attempt was made to generate future synthesized fundus images that can predict AMD progression. In this paper, we developed a deep learning model using fundus images for AMD patients with different time elapses to generate synthetic future fundus images.

Method: We exploit generative adversarial networks (GANs) with additional drusen masks to maintain the pathological information. The dataset included 8196 fundus images from 1263 AMD patients. A proposed GAN-based model, called Multi-Modal GAN (MuMo-GAN), was trained to generate synthetic predicted-future fundus images.

Results: The proposed deep learning model indicates that the additional drusen masks can help to learn the AMD progression. Our model can generate future fundus images with appropriate pathological features. The drusen development over time is depicted well. Both qualitative and quantitative experiments show that our model is more efficient to monitor the AMD disease as compared to other studies.

Conclusion: This study could help individualized risk prediction for AMD patients. Compared to existing methods, the experimental results show a significant improvement in terms of tracking the AMD stage in both image-level and pixel-level.

Introduction

Age-related macular degeneration (AMD) is a disease of the central retina [12]. And one of the most common diseases that can cause blind worldwidely and expected to increase exponentially due to the aging population. Like other diseases, predicting the progression of AMD and identifying eyes with the high risk of progression is the crucial step to prevent blindness and minimize the socio-economic loss due to AMD. There are numerous researches on disease course and risk factors of AMD progression, but almost all study results are presented as numeric figures for general AMD population. For example, the 5-year progression of early AMD in Koreans is reported to be about 20% [21]. However, in the real clinical settings, for each AMD patients, the risk of progression differs based on their fundus findings like drusen numbers, size, location and accompanying pigmentary changes. So, for AMD patients, individualized risk estimation and graphic representation for their eyes can be important as much as treatment itself. Nevertheless, individualized risk prediction based on fundus findings AMD still remains an unresolved problem. Thus, we proposed a deep learning-based method in order to make future synthesized fundus images from the current fundus images.

There are several key problems in this task. Firstly, the future synthesized image must be realistic and it has to show accurate changes in comparison with the input image. The output image is supposed to have the same important feautures with the corresponding ground truth images. For AMD disease, the main factor for diagnosis are drusen. However, in the early stage of AMD, drusen are usually small and unclear. Therefore, it is hard to monitor the changes of drusen over time. In addition, the fundus image series of the same patient with different times can be taken by different conditions. It can cause not only the changes of position but also the style of fundus images. Moreover, to train a model to predict future images, we need a dataset that contains a series of images from the same subject at different times (e.g. AFAD dataset [14] for face aging task).

One of the possible approaches for this research is based on deep learning. Nowadays, it is true that deep learning models have outperformed over classical approaches in many tasks: classify/generate image, text, and speech. Especially, Generative Adversarial Networks (GANs) could be used to generate realistic images [4], [11]. GANs is a deep learning model for learning the distribution of data. The Conditional GANs (cGANs) [7] is a variant of GANs. It takes an image as an input instead of a random vector and then generates an output image. cGANs is popular in many applications such as image segmentation [19], style transfer [22], image-to-image translation [7], etc. Therefore, in this paper, we proposed a GAN-based model to generate future fundus images for AMD patients. To the best of our knowledge, this study is the first attempt to generate future fundus images based on current fundus images and the first attempt to develop an individualized and visually represented risk prediction algorithm. Overall, our contributions in this research are as follows:

  • We introduced a novel deep learning framework that can generate synthetic predicted-future fundus images for early AMD patients. Given the series of fundus images, our model predicts the future fundus image. By capturing AMD factors, the proposed model can provide helpful information of AMD progression and can be used for monitoring the disease.

  • We proposed a new GAN model for this task. We combine GAN and a segmentation model to extract and analyze the AMD factors. Our model exploits drusen masks along with image data to impose important features. Moreover, an extra discriminator is applied for preserving identity.

  • We develop a fundus dataset that fits into our task. The dataset contains fundus images of the 1263 AMD patients. Each patient has several fundus images taken at various times.

Section snippets

Related works

Until now, the most important features to monitor AMD progression are changes in fundus images and Optical Coherence Tomography (OCT) findings. Although there are recent studies regarding OCT images to predict early AMD conversion to wet age-related macular degeneration using deep learning [27], as same as previous risk estimation studies, most of these study results are presented as numeric figures and general estimates. Moreover, acquiring fundus images is more accessible and much cheaper

Image preprocessing

Our pre-processing steps are illustrated in Fig. 1. The important of fundus image for AMD diagnosis is the center region. Therefore, we decide to use only that region for experiments. In order to crop images, the first step is to locate the optic disc position by using the pre-trained model called DiscSeg [2], [3]. DiscSeg is the U-shape convolutional network trained to predict the position of optic disc at the given fundus image. After that, the region of interest (ROI) is cropped based on the

The experiment settings

The dataset includes 8196 of fundus images from 1263 AMD patients. The age of patients are from 30 to 84 years-old. We use 7156 images for training and 1040 images for testing. The details of our dataset are shown in Table 1. For training the main model, we use learning rate of 0.0002 and Adam optimizer with β1=0.0 and β2=0.9. The model is trained with 200 epochs and a batch size of 32. The augmentation techniques are also applied (horizontal flip, shift, scale, and rotate). The Pix2pix model

Conclusions

In this study, we introduce a deep learning model for generating future fundus images of early AMD patients. Our GAN model with two discriminators shows that it can learn and produce the realistic fundus images. By exploiting drusen segmentation masks, we guide the model to focus on drusen changes which is the key factor for diagnosis AMD. The experimental results indicate that our method is able to catch the drusen changes over time.

Since this is the first study for generating future fundus

CRediT authorship contribution statement

Quang T.M. Pham: Conceptualization, Methodology, Software, Validation, Formal analysis, Writing – original draft, Writing – review & editing. Sangil Ahn: Methodology, Formal analysis, Writing – review & editing. Jitae Shin: Conceptualization, Formal analysis, Writing – review & editing, Supervision, Project administration, Funding acquisition. Su Jeong Song: Conceptualization, Data curation, Writing – review & editing.

Declaration of Competing Interest

The authors declare no conflict of interest.

Acknowledgement

This work was partly supported by the National Research Foundation of Korea(NRF) grant funded bythe Korea government(MSIT) (No. 2020R1F1A1065626) and was partly supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2021-2018-0-01798) supervised by the IITP (Institute for Information & communications Technology Promotion). It was also partly supported by the research fund from Biomedical Institute for Convergence

References (27)

  • L.S. Lim et al.

    Age-related macular degeneration

    The Lancet

    (2012)
  • Y. Peng et al.

    Deepseenet: a deep learning model for automated classification of patient-based age-related macular degeneration severity from color fundus photographs

    Ophthalmology

    (2019)
  • L.-C. Chen et al.

    Encoder-decoder with atrous separable convolution for semantic image segmentation

  • H. Fu et al.

    Joint optic disc and cup segmentation based on multi-label deep network and polar transformation

    IEEE Trans Med Imaging

    (2018)
  • H. Fu et al.

    Disc-aware ensemble network for glaucoma screening from fundus image

    IEEE Trans Med Imaging

    (2018)
  • I.J. Goodfellow et al.

    Generative adversarial nets

    Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2

    (2014)
  • K. He et al.

    Deep residual learning for image recognition

    2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)

    (2016)
  • M. Heusel et al.

    Gans trained by a two time-scale update rule converge to a local nash equilibrium

    Proceedings of the 31st International Conference on Neural Information Processing Systems

    (2017)
  • P. Isola et al.

    Image-to-image translation with conditional adversarial networks

    2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)

    (2017)
  • S.A. Kamran et al.

    Fundus2angio: A conditional gan architecture for generating fluorescein angiography images from retinal fundus photography

  • S.A. Kamran et al.

    Attention2angiogan: Synthesizing fluorescein angiography from retinal fundus images using generative adversarial networks

    2020 25th International Conference on Pattern Recognition (ICPR)

    (2021)
  • S.A. Kamran, K.F. Hossain, A. Tavakkoli, S.L. Zuckerbrod, S.A. Baker, Vtgan: Semi-supervised retinal image synthesis...
  • T. Karras, S. Laine, T. Aila, A Style-Based Generator Architecture for Generative Adversarial Networks (2018)....
  • Cited by (17)

    • Vision Transformers in medical computer vision—A contemplative retrospection

      2023, Engineering Applications of Artificial Intelligence
    View all citing articles on Scopus
    View full text