Generating future fundus images for early age-related macular degeneration based on generative adversarial networks
Introduction
Age-related macular degeneration (AMD) is a disease of the central retina [12]. And one of the most common diseases that can cause blind worldwidely and expected to increase exponentially due to the aging population. Like other diseases, predicting the progression of AMD and identifying eyes with the high risk of progression is the crucial step to prevent blindness and minimize the socio-economic loss due to AMD. There are numerous researches on disease course and risk factors of AMD progression, but almost all study results are presented as numeric figures for general AMD population. For example, the 5-year progression of early AMD in Koreans is reported to be about 20% [21]. However, in the real clinical settings, for each AMD patients, the risk of progression differs based on their fundus findings like drusen numbers, size, location and accompanying pigmentary changes. So, for AMD patients, individualized risk estimation and graphic representation for their eyes can be important as much as treatment itself. Nevertheless, individualized risk prediction based on fundus findings AMD still remains an unresolved problem. Thus, we proposed a deep learning-based method in order to make future synthesized fundus images from the current fundus images.
There are several key problems in this task. Firstly, the future synthesized image must be realistic and it has to show accurate changes in comparison with the input image. The output image is supposed to have the same important feautures with the corresponding ground truth images. For AMD disease, the main factor for diagnosis are drusen. However, in the early stage of AMD, drusen are usually small and unclear. Therefore, it is hard to monitor the changes of drusen over time. In addition, the fundus image series of the same patient with different times can be taken by different conditions. It can cause not only the changes of position but also the style of fundus images. Moreover, to train a model to predict future images, we need a dataset that contains a series of images from the same subject at different times (e.g. AFAD dataset [14] for face aging task).
One of the possible approaches for this research is based on deep learning. Nowadays, it is true that deep learning models have outperformed over classical approaches in many tasks: classify/generate image, text, and speech. Especially, Generative Adversarial Networks (GANs) could be used to generate realistic images [4], [11]. GANs is a deep learning model for learning the distribution of data. The Conditional GANs (cGANs) [7] is a variant of GANs. It takes an image as an input instead of a random vector and then generates an output image. cGANs is popular in many applications such as image segmentation [19], style transfer [22], image-to-image translation [7], etc. Therefore, in this paper, we proposed a GAN-based model to generate future fundus images for AMD patients. To the best of our knowledge, this study is the first attempt to generate future fundus images based on current fundus images and the first attempt to develop an individualized and visually represented risk prediction algorithm. Overall, our contributions in this research are as follows:
- •
We introduced a novel deep learning framework that can generate synthetic predicted-future fundus images for early AMD patients. Given the series of fundus images, our model predicts the future fundus image. By capturing AMD factors, the proposed model can provide helpful information of AMD progression and can be used for monitoring the disease.
- •
We proposed a new GAN model for this task. We combine GAN and a segmentation model to extract and analyze the AMD factors. Our model exploits drusen masks along with image data to impose important features. Moreover, an extra discriminator is applied for preserving identity.
- •
We develop a fundus dataset that fits into our task. The dataset contains fundus images of the 1263 AMD patients. Each patient has several fundus images taken at various times.
Section snippets
Related works
Until now, the most important features to monitor AMD progression are changes in fundus images and Optical Coherence Tomography (OCT) findings. Although there are recent studies regarding OCT images to predict early AMD conversion to wet age-related macular degeneration using deep learning [27], as same as previous risk estimation studies, most of these study results are presented as numeric figures and general estimates. Moreover, acquiring fundus images is more accessible and much cheaper
Image preprocessing
Our pre-processing steps are illustrated in Fig. 1. The important of fundus image for AMD diagnosis is the center region. Therefore, we decide to use only that region for experiments. In order to crop images, the first step is to locate the optic disc position by using the pre-trained model called DiscSeg [2], [3]. DiscSeg is the U-shape convolutional network trained to predict the position of optic disc at the given fundus image. After that, the region of interest (ROI) is cropped based on the
The experiment settings
The dataset includes 8196 of fundus images from 1263 AMD patients. The age of patients are from 30 to 84 years-old. We use 7156 images for training and 1040 images for testing. The details of our dataset are shown in Table 1. For training the main model, we use learning rate of 0.0002 and Adam optimizer with and . The model is trained with 200 epochs and a batch size of 32. The augmentation techniques are also applied (horizontal flip, shift, scale, and rotate). The Pix2pix model
Conclusions
In this study, we introduce a deep learning model for generating future fundus images of early AMD patients. Our GAN model with two discriminators shows that it can learn and produce the realistic fundus images. By exploiting drusen segmentation masks, we guide the model to focus on drusen changes which is the key factor for diagnosis AMD. The experimental results indicate that our method is able to catch the drusen changes over time.
Since this is the first study for generating future fundus
CRediT authorship contribution statement
Quang T.M. Pham: Conceptualization, Methodology, Software, Validation, Formal analysis, Writing – original draft, Writing – review & editing. Sangil Ahn: Methodology, Formal analysis, Writing – review & editing. Jitae Shin: Conceptualization, Formal analysis, Writing – review & editing, Supervision, Project administration, Funding acquisition. Su Jeong Song: Conceptualization, Data curation, Writing – review & editing.
Declaration of Competing Interest
The authors declare no conflict of interest.
Acknowledgement
This work was partly supported by the National Research Foundation of Korea(NRF) grant funded bythe Korea government(MSIT) (No. 2020R1F1A1065626) and was partly supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2021-2018-0-01798) supervised by the IITP (Institute for Information & communications Technology Promotion). It was also partly supported by the research fund from Biomedical Institute for Convergence
References (27)
- et al.
Age-related macular degeneration
The Lancet
(2012) - et al.
Deepseenet: a deep learning model for automated classification of patient-based age-related macular degeneration severity from color fundus photographs
Ophthalmology
(2019) - et al.
Encoder-decoder with atrous separable convolution for semantic image segmentation
- et al.
Joint optic disc and cup segmentation based on multi-label deep network and polar transformation
IEEE Trans Med Imaging
(2018) - et al.
Disc-aware ensemble network for glaucoma screening from fundus image
IEEE Trans Med Imaging
(2018) - et al.
Generative adversarial nets
Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2
(2014) - et al.
Deep residual learning for image recognition
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
(2016) - et al.
Gans trained by a two time-scale update rule converge to a local nash equilibrium
Proceedings of the 31st International Conference on Neural Information Processing Systems
(2017) - et al.
Image-to-image translation with conditional adversarial networks
2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
(2017) - et al.
Fundus2angio: A conditional gan architecture for generating fluorescein angiography images from retinal fundus photography
Attention2angiogan: Synthesizing fluorescein angiography from retinal fundus images using generative adversarial networks
2020 25th International Conference on Pattern Recognition (ICPR)
Cited by (17)
Increasing-Margin Adversarial (IMA) training to improve adversarial robustness of neural networks
2023, Computer Methods and Programs in BiomedicineVision Transformers in medical computer vision—A contemplative retrospection
2023, Engineering Applications of Artificial IntelligenceLearn Single-horizon Disease Evolution for Predictive Generation of Post-therapeutic Neovascular Age-related Macular Degeneration
2023, Computer Methods and Programs in BiomedicineA GAN-based deep enhancer for quality enhancement of retinal images photographed by a handheld fundus camera
2022, Advances in Ophthalmology Practice and Research