Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Diagnosis of retinal disorders from Optical Coherence Tomography images using CNN

  • Nithya Rajagopalan ,

    Roles Conceptualization, Formal analysis, Investigation, Methodology, Software, Validation, Visualization, Writing – original draft

    nithyar@ssn.edu.in

    Affiliation Department of Biomedical Engineering, Sri Sivasubramaniya Nadar College of Engineering, Chennai, India

  • Venkateswaran N.,

    Roles Investigation, Supervision, Writing – review & editing

    Affiliation Department of Electronics and Communication Engineering, Sri Sivasubramaniya Nadar College of Engineering, Chennai, India

  • Alex Noel Josephraj,

    Roles Methodology, Software, Supervision, Visualization, Writing – review & editing

    Affiliation Department of Electronic Engineering, College of Engineering, Shantou University, Shantou, China

  • Srithaladevi E.

    Roles Data curation, Investigation, Methodology, Software

    Affiliation Department of Biomedical Engineering, Sri Sivasubramaniya Nadar College of Engineering, Chennai, India

Abstract

An efficient automatic decision support system for detection of retinal disorders is important and is the need of the hour. Optical Coherence Tomography (OCT) is the current imaging modality for the early detection of retinal disorders non-invasively. In this work, a Convolution Neural Network (CNN) model is proposed to classify three types of retinal disorders namely: Choroidal neovascularization (CNV), Drusen macular degeneration (DMD) and Diabetic macular edema (DME). The hyperparameters of the model like batch size, number of epochs, dropout rate, and the type of optimizer are tuned using random search optimization method for better performance to classify different retinal disorders. The proposed architecture provides an accuracy of 97.01%, sensitivity of 93.43%, and 98.07% specificity and it outperformed other existing models, when compared. The proposed model can be used for the large-scale screening of retinal disorders effectively.

1. Introduction

The eye is the light of human life. The light enters the eye through the cornea, passing through the aqueous humor, lens, vitreous humor, and finally on to the retina lying on the back of eye [1]. The retina is the most important part of the eye. It is divided into pigment epithelium, receptor layer, cell layer, receptor layer, internal limiting membrane, external limiting membrane, and vitreous body. The retina’s photoreceptor cells use the external light that the cornea focuses through the lens and convert it into nerve signals, transported to the brain through the optic nerve.

According to WHO’s Blindness and vision impairment statistics in 2019, around 2.2 billion people suffer from blindness or vision impairment, out of which 80% of disability can be avoided if detected at an early stage [1, 2]. Around 6.9 million people have glaucoma, 3 million people have diabetic retinopathy, and 2.75 million people have AMD and Duren. Also, WHO (2019) estimates that more than 360 million people will be affected worldwide by diabetes mellitus by 2030. All these people will be at risk of developing diabetic macular edema [3].

With such massive numbers, detecting retinal disorders manually by ophthalmologist is a strenuous task. The identification of retinal disorders can be made by spotting the existence of deformation associated with the disease. Although clinicians may be able to diagnose the disease through vascular abnormalities, its resource demands are high. In underdeveloped countries where the diabetic population is high, this equipment may not be readily accessible. The major challenges are mass screening, skilled technician, observer variability and, early detection. An automated alternative is more efficient, reliable, and the need of the hour. The main contribution of the work is to develop an algorithm for detecting retinal disorders towards building an efficient decision support system.

Diabetic Macular Edema (DME) causes retinal inflammation and leakage of blood vessels due to Diabetic Retinopathy (DR) [4]. The prevalence of DME in DR subjects is 2.7% - 11%. Common factors affecting DME occurrence are ethnicity, gender, proteinuria, cardiovascular diseases, and Diabetes. In Drusen Macular Degeneration (DMD), the occurrence of yellow or white deposits between the sub retinal layers were more common. Risk of macular degeneration is high with the progression of such deposits leading to vision loss [57]. In Choroid Neovascularization (CNV) condition, the non-vascularized blood vessel enters the RPE cells and causes vascular leakage. In CNV, it is observed that new blood vessels emerge from the choroidal region, which often causes hemorrhage. The advanced stage of CNV results in the thickening of the Retinal Pigment Epithelium (RPE) layer of the retina. Detecting retinal disorders at an early stage is used to prevent disease progression and vision loss [8].

An ophthalmologist uses two major imaging modalities, namely color fundus imaging and OCT. The 2D image of the retina is well represented in the color fundus image. The fundus camera captures the light reflected from the retina and forms the fundus image. In a color fundus image, deformation in the retina can be identified. But the depth information degeneration cannot be accessed. The OCT is a non-invasive imaging modality mainly used in ophthalmology to visualize retinal layers [9, 10]. Information about all the retina layers can be inferred from OCT images, useful in detecting and diagnosing retinal disorders [8, 10]. Early detection of retinal disorders can be done effectively using OCT compared to fundus photography. Even a minimal change in retinal layers can be accurately seen in the OCT images, as it acquires the cross-sectional view from the sub retinal layers. Fundus photography can be used only for visualizing the 2-dimensional view of the retina and lacks providing depth information about the retinal layer [11, 12]. The interferometric technique’s properties are defined by the signal sampling at the detector and the light source’s coherence properties. With this unique OCT property, the retina’s high-resolution image is achieved. Timed domain OCT and Frequency domain OCT are different type of acquisition domains. The light source used in TD-OCT is usually a super luminescent diode and reference beam length is varied. Frequency domain OCT (FD-OCT) uses separate detectors to acquire the broadband interference.

In this work the author develops a CNN architecture for detection of retinal disorders using OCT images. The rest of the paper is organized as follows. Related work is discussed in section II. The results and discussion for the proposed architecture is provided in section IV. Section V subsumes the conclusion and course of action for the future.

2. Related works

Retinal disorders are detected and diagnosed by performing retinal layer segmentation and thickness measurement in the retinal OCT image. The retinal layers’ changes due to any disorder were not common, and fixing a specific benchmark process is impossible for analyzing the data. Standard image processing algorithms for retinal layer abnormality detection have some difficulties, such as time-consuming, sufficient domain knowledge. Also, a generalization of the process for automatic processing is difficult [13, 14]. A convolution neural network is a recent tool that involves image classification, image recognition, and image retrieval [1517].

Often-occurring retinal complications, which involve damage to the optic disc, macular region, rods, cones, and blood vessels, supply and nourish the retina, resulting in vision loss. For the detection and treatment of retinal disorders, machine learning algorithms have been widely employed. Machine learning algorithms yield a function that can accurately predict class labels based on a training algorithm. The retinal image is processed through a sequential process of image pre-processing, segmentation, feature extraction, supervised and unsupervised classification methods to detect various retinal disorders [18]. For automated DR detection, [19] proposed a transfer learning-based CNN on binocular retinal fundus images. A hybrid deep learning model to detect retinal lesions automatically is presented [20]. A CNN algorithm for DR grading. With less volume of dataset, the performance decreased during testing [21].

CNN can effectively extract features and classify retinal OCT images [22, 23]. Rare retinal disorders like inferior staphyloma, chorioretinal atrophy, Vogt—Koyanagi—Harada (VKH) the disease can also be effectively detected using the CNN model [24]. Several CNN networks are framed for various applications. The LeNet network has few free parameters, and also, the network could be trained on a low-level representation of data that has minimal pre-processing [25]. Another network called AlexNet has improved performance over LeNet architecture. AlexNet was the first deep network architecture used for Image classification to classify a thousand classes [26, 27]. This network is used for the DR detection using retinal fundus images. Three stages of DR have been classified with an average accuracy of 96% [28]. The VGG-16 architecture has been proposed for obtaining high-level features [29, 30] and the number of parameters used in the architecture is 138 million parameters. The complexity of VGG-16 is high compared to the Alexnet [31]. With the CNN model, image classification [32] and disease diagnosis can be made efficiently with less processing time [17, 33, 34]. Most of the literature concentrates on fundus image for retinal disorder detection which gives only two-dimensional information [18, 20, 21]. The OCT image is able to detect the disease at an early stage, with the available three-dimensional information of the retinal layers. Recently researchers have concentrated on OCT images. Noise removal and disease diagnosis is a challenging task in OCT images [14, 35, 36]. A simple network with high efficiency is developed to detect retinal disorders using OCT images.

Analysis on noise removal is implemented to remove the speckle noise in the OCT images for improving the efficiency of the system. Development of a simple CNN model for four class classification of retinal disorders and tuning its hyperparameters using random search optimization method. In most of the CNN models, a transfer learning approach is used that modifies the existing architecture for necessary applications to achieve better efficiency. In this paper, a novel CNN model is proposed having less complexity, low computational time and tuned hyperparameters. Also, in this work a four-class classification is performed for retinal disorders detection.

This research aims to classify retinal disorders using a novel CNN model with higher accuracy than the existing model. The features of convolution layers are also visualized. The OCT retinal images of normal and abnormal conditions are shown in Fig 1(A)–1(D).

thumbnail
Fig 1.

Retinal OCT images of (a) Normal (b) CNV (c) DMD (d) DME.

https://doi.org/10.1371/journal.pone.0254180.g001

The Mendeley database consists of around 84,000 OCT images are categorized into four types namely Diabetic Macular Edema, Drusen Macular Degeneration, Choroidal Neovascularization and Normal [35]. The retinal OCT of CNV, DME, DMD and normal images were used in this study. From the database, 12000 images have been used in the classification process, 3000 images from each category.

3. Pre-processing of OCT images

Pre-processing the images is to improve the quality of the image so that interpretation without error is possible, and enhance some image features that are important for further processing. Generally, OCT images have speckle noise, a granular noise commonly found in medical images. It is a multiplicative expressed as shown in Eq 1.

(1)

Where c(i,j) is noisy image, o(i,j) original image, u(i,j) is the multiplicative parameter, and η(i,j) is the additive parameter.

This type of speckle noise reduction improves the visual perception and system accuracy during classification. The Retinal OCT images are to be pre-processed for removing speckle noise. Firstly, the images are resized to equal dimensions since they are all of the different sizes. Secondly, in the speckle noise reduction process, different filters are used to compare performance factors. Both spatial domain–convolving the noisy image with mask, and frequency domain filters–transforming the noisy image to frequency domain and apply filters are used for speckle-noise reduction. The spatial filter acts on an image by reducing the intensity variation between adjacent pixels. The simple sliding-window spatial filter replaces itself and pixels that are unrepresentative of their surroundings. It is implemented with a convolution square mask that provides a result which is a weighted sum of the values of a pixel and its neighbors. If the mask’s sum coefficients are to unit value, then the average brightness of the image is not changed, however while the sum gives zero, the average brightness is lost, and it returns a dark image. The common speckle filters such as Kuan, mean filter, biorthogonal spline wavelet, and wiener filter are considered for this study.

Kuan filter is an additive smoothening filter that changes the multiplicative speckle parameter into an additive linear parameter. It is often used to remove speckle noise from the radar and medical images using minimum mean square error calculation to estimate the signal’s value for the center cell in the window. It then calculates the signal estimate from the variance, local mean, and standard deviation. The weighted function is expressed as in Eq 2.

(2)

Where Cu is the estimated noise variation coefficient, and it is expressed as: Cu = , ENL is equivalent noise looks.

Ci = S/m - Variation coefficient of image

S = Standard deviation in the filter window, m = mean with in the filter window

Mean filters are simple and intuitive filters which reduce the amount of variation in intensity between one pixel and the next. Mean filter simply replaces each pixel value in an image with its neighbors’ mean value, including itself. By eliminating pixel values that are unrepresentative of surroundings, the speckle noise is suppressed. It computes the sum of all pixels in the sliding window and then divides the sum by the number of pixels in the filter window. The result gives a blurring effect with some loss of details along with reduced speckle noise.

For a*b window region, a mathematical representation of the mean filter is given as in the Eq 3.

(3)

The biorthogonal wavelets introduced by Cohen-Daubechies-Feauveau (CDF) wavelets are arguably the second most popular family of wavelets in image processing. A variant of these with four primal and dual vanishing moments is used in the JPEG2000 standard. They are also popular in finite element computations in scientific computing because the elementary scaling functions can be B-spline. This means they are piecewise polynomials, and they can be represented exactly with analytical expressions as in Eq 4.

(4)

The Wiener filter is a linear spatial domain filter. It can restore corrupted or blurred images. It can also be implemented in both spatial domain (mean squared method for denoising) and frequency domain (Fourier transforms for denoising and blurring operation). It works on the basis of computation of specific statistical parameters apart from usual parameters (mean, Variance), both the locally (higher order moments of the kernel) and globally (higher-order moment of the entire image), and the statistical properties in the image differ from one region to another. The larger the local variance, the lesser the smoothing effect, and if the local variance is small. It is represented mathematically as in Eq 5.

(5)

Where,

H(u, v)2 = Degradation function and

H(u, v)* = Conjugate complex

w(u, v) = Degraded image

sn(u, v) = Power spectra of noise

sf(u, v) = Power spectra of original image

To analyze above mentioned filtering techniques three parameters namely, Peak Signal-to-Noise Ratio (PSNR), Mean Square Error (MSE), and structural similarity index (SSIM) are calculated. The noisy image and filtered image are considered for performance analysis of the filter. When the PSNR values are low, with high MSE and SSIM value, the filter performance is high.

Mean Square Error (MSE) is the cumulative squared error between the filtered and the original image it expressed as in Eq 6.

(6)

Where,

I1(i,j) = Original image

I2(i,j) = Approximated version of the image (filtered image) and A,B = Dimensions of the images

Peak Signal-to-Noise Ratio (PSNR) is the ratio between the maximum possible power and the power of noise. It is expressed as in Eq 7.

(7)

Where N = representing the number of gray levels. Structural similarity Index (SSIM) is a perceptual metric that measures image quantity degradation caused by compression due to processing or loss in data transmission. It is also a full reference metric requiring both the original image and the measurement’s processed image. SSIM can be defined as in Eq 8.

(8)

Where μ1, μ2 - the mean value of the original and the filtered image

σ1, σ2 - the standard deviations of the original and the filtered image.

σ1,2- the covariance between the original and the filtered image.

The parameter value varies between 0 and 1, and low value represents structurally dissimilar. The images in the dataset are of different sizes. Therefore, all the images are resized to equal size to have an equal number of pixels. Four different filters are used for speckle noise reduction. The filters are, mean filter, wiener filter, Kuan and biorthogonal spline wavelet filter. For analyzing the speckle reduction, the quality assessment metrics are calculated. Original and filtered images are considered for the measurement. The PSNR, MSE, and SSIM of the various filters used are represented in the Table 1, from that it is inferred biorthogonal spline wavelet based filter outperformed the other three. Fig 2 shows the output of filters applied on the OCT images. The input images were preprocessed and then fed to the proposed CNN Model for classification.

thumbnail
Fig 2.

Output of the Speckle Reduction Filters (a) Original image, (b) Biorthogonal Spline wavelet (c) Mean filter, (d) Kuan filter, and (e) Wiener filter.

https://doi.org/10.1371/journal.pone.0254180.g002

4. Convolution neural network

Some basic components required for framing CNN are convolution layers (CL), pooling layers (PL), and a fully connected layer (FCL). The convolution layer uses the activation functions like ReLU, softmax etc., followed by the pooling layer. This pattern was repeated until the image is merged spatially to reduce its size. Following this, it was connected to a fully connected layer (FCL). The output can be obtained from FCL. These layers were stacked to form the full CNN architecture. The CNN components are described below. The CL specification can be defined by parameters that include filter size, strides, padding, and spatial size (width, height, and depth). Each filter in the layer slides over the input volume’s spatial size and calculates the dot product between the filter and input volume elements. The results obtained were passed into the non-linear activation function. ReLU is the activation function followed by each CL. ReLU function is the most commonly used activation, which works much faster than other activation functions like tanh or sigmoidal function [32]. The PL is inserted in between CL in the architecture. The PL’s function is to make a progressive reduction in the size of the input and for the computation of parameters. The fully connected layer (FCL) is the last layer of the convolution architecture. The function of FCL is to connect all neurons from the former layer to the single neuron layer. The softmax activation function is the most commonly used in the final layer of CNN because of its better probability distribution.

4.1 Proposed CNN architecture

The proposed CNN Architecture contains five Convolution Layers (CL) and two fully connected Layers (FL). The CNN components were described below. The CL specification can be defined by parameters that include filter size, strides, padding, and spatial size (width, height, and depth). Each filter in the layer slides over the input volume’s spatial size and calculates the dot product between the filter and input volume elements. The results obtained were passed into the non-linear activation function. ReLU is the activation function followed by each CL. The PL is inserted in between CL in the architecture. The PL’s function is to make a progressive reduction in the size of the input and computation of parameters. The fully connected layer (FCL) is the last layer of the convolution architecture. The function of FCL is to connect all neurons from the former layer to the single neuron layer. The original images were in different dimensions. The images were resized into uniform size to enable extraction of the features. Filtered for speckle noise reduction. The input image was fed into the network with a spatial dimension of 224x224x1. For this work, 12,000 retinal greyscale images were considered, 3,000 images in each of four categories. The dataset was split for training and testing. Using the K- fold method, the images were split into 8000 images for training and 4,000 for testing. The validation split was 0.1 from the training dataset. Less complexity and hyperparameter optimization are the advantage of the proposed architecture. The computation time to train the network takes 4s for each epoch. The model is implemented with a 32 core AMD processor, 64GB RAM with NVidia 2060RTX series GPU.

The resized images were passed through the stack of CLs. The ReLU activation function was used in the CLs, which converged faster than other activation functions. The max-pooling layer was followed by the CLs (not all convolution layers) with a common pool size of (2,2) with strides of 2. Two FLCs followed the CLs. The output layer was the softmax activation layer, which classified the input into four classes. The stochastic gradient descent (SGD) optimizer was used to reduce the error rate and metrics like accuracy and loss function. Overfitting the model can be avoided by choosing the optimal dropout rate [24, 37].

Batch Normalization is a method commonly used in CNNs to normalize a set of inputs to the layer. The input layer can be normalized by adjusting and scaling the activations. It has several advantages, namely, its ability to reduce overfitting, activation adjustment, and its mutual relationship with the dropout layer. In the proposed architecture, the batch normalization is incorporated in the Convolutional and Dense layers. The proposed network architecture is shown in Fig 3.

The output size and trainable parameters of proposed CNN layers were computed using Eqs 9 and 10.

(9)(10)

The details about the layers in the proposed CNN model are given in Table 2.

4.2 Hyperparameters optimization

The critical task of deep learning is to choose the best hyperparameters for the model. The hyperparameter selection by manual search involves many attempts and costs serious. So, the authors moved on to optimize the hyperparameters. The hyperparameters used in deep learning were learning rate, batch size, epochs, and optimizer. There are several approaches for hyperparameter tuning. The most widely used hyperparameter tuning technique is the random search. Random search is a technique that selects a combination of hyperparameters for training the model [38]. In this investigation, a random search was performed. The hyperparameters were tuned for batch size, epochs, dropout rate, and optimizer. The hyperparameters were evaluated using mean and standard deviation. The optimized hyperparameters are shown in Table 3.

4.3 Feature visualization

The performance of each convolution layer can be seen by visualizing the output of each layer. Feature visualization explicitly shows the functionality of each CNN layer [39]. Initial layers extract the edge information by neighbor comparison and advanced layers extract higher level features used for classification. Fig 4 visualizations output of the first convolution layer (64 filters), highlighting the edge information.

thumbnail
Fig 4. Visualization of features of first convolution layers.

https://doi.org/10.1371/journal.pone.0254180.g004

5. Experimental results and discussion

The network is trained with 8000 images, where 2000 were trained in each of the four classes. The proposed network architecture’s accuracy and loss curve is shown in Fig 5(A) & 5(B). The validation split took 0.1% of the data from the four classes. The proposed network achieved validation accuracy was 98.4%, whereas the training loss was reduced to 25% and validation loss to 10%.

5.1 Choice of K value

K-fold cross-validation helps improve the efficiency of the customized model. The k-fold divides the data into k groups and gets trained by each data. Therefore, it predicts easily when the model explores unseen data with the same features as the trained images. In this work, the k values set were assigned to 3, 4, and 5. This provided the ability to measure the proposed model’s performance to achieve a higher testing accuracy with three folds. With the increase in k value, the accuracy decreases, as shown in Table 4.

The above result shows the parameters like average testing accuracy of the proposed model reaching a higher accuracy for threefold.

5.2 Confusion matrix for four classes (3-Fold)

The customized model performed a multi-class classification. The network was tested with 4000 images, with 1000 drawn from each of the four categories. The classifier distinguished between urgent referrals like CNV, DME, and DMD from normal.

Table 5 shows that the classifier was correctly predicting 961 images as CNV, 923 images DME, 926 images as DMD, and 933 images as normal. Prediction of the remaining cells in the confusion table was erroneous. Performance metrics were calculated from the confusion matrix and tabulated in Table 6.

5.3 Performance comparison

The proposed convolution neural network was compared with existing architectures. The dataset was trained and tested with LeNet and AlexNet. The architectural complexity is less in these two networks similar to the developed model. So, these models were trained and tested with the same dataset. The performance of the proposed model is compared and found outperforming as in Table 7.

The receiver operating characteristic (ROC) curve is a graphical plot that shows the classifier’s diagnostic ability. The performance of the classifier was better when the value was higher. If the value is higher, then the performance of the classifier is better. The ROC curve plots the true positive rate (TPR) against the false positive rate (FPR) at different thresholds. The ROC curve of the proposed architecture is shown in Fig 6.

The ROC rate for class 0 is 99%, ROC rate for class 1 is 96%, ROC rate for class 2 is 96%, and ROC rate for class 3 is 98%. The ROC curve of AlexNet and LeNet architecture, as shown in Fig 7(A) & 7(B), were obtained and used for comparison with of ROC curve of the proposed network architecture.

The database provider [35] used transfer learning techniques in ImgeNet and achieved an accuracy of 93.4% with a sensitivity and specificity of 96.6% and 94% respectively. The area under the ROC curve classifies the disorders from normal is 98.8%. The accuracy of the proposed model when compared to the above model is better.

The results obtained were compared with the existing research papers and presented in Table 8.

The OCT image denoising analysis is performed and found Biorthogonal wavelet transform is better in denoising the image. The denoised image is fed to the proposed CNN model for classification. The results show the proposed model has higher efficiency in the classification of OCT images into disease categories such as CNV, DMD, DME, and Normal. The performance of the network is explained by highlighting the confusion matrix, calculating the average accuracy, sensitivity and specificity. The results of the proposed model are compared with few of the literature and existing models.

6. Conclusion

A four-class classification is performed for retinal OCT images using proposed CNN architecture. From the publicly available MENDELY dataset, four classes of images, CNV, DMD, DME, and normal (3000 in each class). The images were denoised using a biorthogonal spline wavelet filter. The denoised images were fed to an eleven-layer CNN architecture framed to classify the retinal disorders. A random search method was used to optimize the hyperparameters. This method can be used for large-scale screening of retinal disorders effectively. The CNN architecture effectively classified urgent referrals like CNV, DME, and DMD from the normal retinal OCT images with a testing accuracy of 97.01%. Effective visualization of the features of the convolution layers is possible. The proposed network has achieved a sensitivity of 93.43% and a specificity of 98.07%. The network proposed in the paper has been optimized for its hyperparameters. The proposed network has been shown to perform better compared to existing models. The proposed techniques’ demonstrated efficacy could help the ophthalmologists in the effective retinal image analysis, thereby providing a better treatment at an early stage, thus preventing blindness. The proposed model is trained and tested only with OCT images of different retinal disorders. The work can be further extended to develop a single flexible CNN architecture that can analyze different modalities of retinal images like ultrasound, fundus images for different disorders. Such a system can be useful in large-scale screening. For analysis, 12000 images are considered, 3000 images in each category CNV, DMD, DME and normal. In future, the dataset size could be increased for enhancing the efficiency of the system. Such a system can be useful in large-scale screening.

References

  1. 1. Oh S. L. et al., “A deep learning approach for Parkinson’s disease diagnosis from EEG signals,” Neural Comput. Appl., pp. 1–7, Aug. 2018,
  2. 2. Samagaio G., Estévez A., de Moura J., Novo J., Fernández M. I., and Ortega M., “Automatic macular edema identification and characterization using OCT images,” Comput. Methods Programs Biomed., vol. 163, pp. 47–63, 2018, pmid:30119857
  3. 3. Wu L., “Classification of diabetic retinopathy and diabetic macular edema,” World J. Diabetes, vol. 4, no. 6, p. 290, 2013, pmid:24379919
  4. 4. Wenick A. and Bressler N., “Diabetic macular edema: Current and emerging therapies,” Middle East African Journal of Ophthalmology, vol. 19, no. 1. Wolters Kluwer—Medknow Publications, pp. 4–12, Jan. 2012, https://doi.org/10.4103/0974-9233.92110 pmid:22346109
  5. 5. Ambati J. and Fowler B. J., “Mechanisms of age-related macular degeneration,” Neuron, vol. 75, no. 1. NIH Public Access, pp. 26–39, Jul. 12, 2012, pmid:22794258
  6. 6. Schlegl T. et al., “Fully Automated Detection and Quantification of Macular Fluid in OCT Using Deep Learning,” Ophthalmology, vol. 125, no. 4, pp. 549–558, 2018, pmid:29224926
  7. 7. Lee R., Wong T. Y., and Sabanayagam C., “Epidemiology of diabetic retinopathy, diabetic macular edema and related vision loss,” Eye Vis., vol. 2, no. 1, p. 17, Dec. 2015, pmid:26605370
  8. 8. Gołębiewska J., Brydak-Godowska J., Moneta-Wielgoś J., Turczyńska M., Kęcik D., and Hautz W., “Correlation between Choroidal Neovascularization Shown by OCT Angiography and Choroidal Thickness in Patients with Chronic Central Serous Chorioretinopathy,” J. Ophthalmol., vol. 2017, pp. 1–7, Oct. 2017, pmid:29109866
  9. 9. Huang D. et al., “Optical coherence tomography.,” Science, vol. 254, no. 5035, pp. 1178–81, Nov. 1991, pmid:1957169
  10. 10. Ţălu S.-D., “Optical Coherence Tomography in the Diagnosis and Monitoring of Retinal Diseases,” ISRN Biomed. Imaging, vol. 2013, pp. 1–13, 2013,
  11. 11. Strøm C., Sander B., Larsen N., Larsen M., and Lund-Andersen H., “Diabetic macular edema assessed with optical coherence tomography and stereo fundus photography,” Investig. Ophthalmol. Vis. Sci., vol. 43, no. 1, pp. 241–245, 2002. pmid:11773037
  12. 12. T. Hassan, M. U. Akram, B. Hassan, A. Nasim, and S. A. Bazaz, “Review of OCT and fundus images for detection of Macular Edema,” in 2015 IEEE International Conference on Imaging Systems and Techniques (IST), Sep. 2015, pp. 1–4, 10.1109/IST.2015.7294517.
  13. 13. Lang A. et al., “Retinal layer segmentation of macular OCT images using boundary classification,” Biomed. Opt. Express, vol. 4, no. 7, p. 1133, Jul. 2013, pmid:23847738
  14. 14. Motamedi S. et al., “Normative data and minimally detectable change for inner retinal layer thicknesses using a semi-automated OCT image segmentation pipeline,” Front. Neurol., vol. 10, p. 1117, 2019, pmid:31824393
  15. 15. Rajathi V., Bhavani R. R., and Wiselin Jiji G., “Varicose ulcer(C6) wound image tissue classification using multidimensional convolutional neural networks,” Imaging Sci. J., pp. 1–11, Sep. 2019,
  16. 16. Mao J., Zhong D., Hu Y., Sheng W., Xiao G., and Qu Z., “An image authentication technology based on depth residual network,” Syst. Sci. Control Eng., vol. 6, no. 1, pp. 57–70, Jan. 2018,
  17. 17. Seth S. and Agarwal B., “A hybrid deep learning model for detecting diabetic retinopathy,” J. Stat. Manag. Syst., vol. 21, no. 4, pp. 569–574, Jul. 2018,
  18. 18. Geetharamani R. and Balasubramanian L., “Retinal blood vessel segmentation employing image processing and data mining techniques for computerized retinal image analysis,” Integr. Med. Res., vol. 36, pp. 102–118, 2016,
  19. 19. Zeng X., Chen H., Luo Y., and Ye W., “Automated diabetic retinopathy detection based on binocular siamese-like convolutional neural network,” IEEE Access, vol. 7, pp. 30744–30753, 2019,
  20. 20. Abràmoff M. D. et al., “Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning,” Investig. Ophthalmol. Vis. Sci., vol. 57, no. 13, pp. 5200–5206, Oct. 2016, pmid:27701631
  21. 21. Gulshan V. et al., “Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs,” JAMA—J. Am. Med. Assoc., vol. 316, no. 22, pp. 2402–2410, Dec. 2016, pmid:27898976
  22. 22. K. A. Nugroho, “A Comparison of Handcrafted and Deep Neural Network Feature Extraction for Classifying Optical Coherence Tomography (OCT) Images,” in 2018 2nd International Conference on Informatics and Computational Sciences (ICICoS), Oct. 2018, pp. 1–6, 10.1109/ICICOS.2018.8621687.
  23. 23. Perdomo O., Otalora S., Gonzalez F. A., Meriaudeau F., and Muller H., “OCT-NET: A convolutional network for automatic classification of normal and diabetic macular edema using sd-oct volumes,” in Proceedings—International Symposium on Biomedical Imaging, May 2018, vol. 2018-April, pp. 1423–1426,
  24. 24. Kuwayama S. et al., “Automated Detection of Macular Diseases by Optical Coherence Tomography and Artificial Intelligence Machine Learning of Optical Coherence Tomography Images,” J. Ophthalmol., vol. 2019, pp. 1–7, Apr. 2019, pmid:31093370
  25. 25. Lecun Y., Bottou L., Bengio Y., and Haffner P., “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, no. 11, pp. 2278–2324, 1998,
  26. 26. Jing Sun, Xibiao Cai, Fuming Sun, and J. Zhang, “Scene image classification method based on Alex-Net model,” in 2016 3rd International Conference on Informative and Cybernetics for Computational Social Systems (ICCSS), Aug. 2016, pp. 363–367, 10.1109/ICCSS.2016.7586482.
  27. 27. Lakhani P. and Sundaram B., “Deep Learning at Chest Radiography: Automated Classification of Pulmonary Tuberculosis by Using Convolutional Neural Networks,” Radiology, vol. 284, no. 2, pp. 574–582, Aug. 2017, pmid:28436741
  28. 28. Shanthi T. and Sabeenian R. S., “Modified Alexnet architecture for classification of diabetic retinopathy images,” Comput. Electr. Eng., vol. 76, pp. 56–64, Jun. 2019,
  29. 29. Simonyan K. and Zisserman A., “Very Deep Convolutional Networks for Large-Scale Image Recognition,” Sep. 2015.
  30. 30. Geng L., Zhang S., Tong J., and Xiao Z., “Lung segmentation method with dilated convolution based on VGG-16 network,” Comput. Assist. Surg., pp. 1–7, Aug. 2019, pmid:31402721
  31. 31. Grm K., Štruc V., Artiges A., Caron M., and Ekenel H. K., “Strengths and weaknesses of deep learning models for face recognition against image degradations,” IET Biometrics, vol. 7, no. 1, pp. 81–89, Jan. 2018,
  32. 32. Krizhevsky A., Sutskever I., and Hinton G. E., “ImageNet Classification with Deep Convolutional Neural Networks,” Commuications ACM, vol. 60, no. 6, pp. 84–90, 2017, Accessed: Sep. 26, 2019. [Online]. Available: http://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networ.
  33. 33. Grassmann F. et al., “A Deep Learning Algorithm for Prediction of Age-Related Eye Disease Study Severity Scale for Age-Related Macular Degeneration from Color Fundus Photography,” Ophthalmology, vol. 125, no. 9, pp. 1410–1420, 2018, pmid:29653860
  34. 34. Schmidt-Erfurth U., Sadeghipour A., Gerendas B. S., Waldstein S. M., and Bogunović H., “Artificial intelligence in retina,” Prog. Retin. Eye Res., vol. 67, no. May, pp. 1–29, 2018, pmid:30076935
  35. 35. Kermany D. S. et al., “Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning,” Cell, 2018, pmid:29474911
  36. 36. Yang Q. et al., “Automated layer segmentation of macular OCT images using dual-scale gradient information,” Opt. Express, vol. 18, no. 20, p. 21293, Sep. 2010, pmid:20941025
  37. 37. Srivastava N., Hinton G., Krizhevsky A., Sutskever I., and Salakhutdinov R., “Dropout: A Simple Way to Prevent Neural Networks from Overfitting,” J. Mach. Learn. Res., vol. 15, pp. 1929–1958, 2014, Accessed: Sep. 20, 2019. [Online]. Available: http://jmlr.org/papers/v15/srivastava14a.html.
  38. 38. Bergstra J. and Bengio Y., “Random Search for Hyper-Parameter Optimization Yoshua Bengio,” J. Mach. Learn. Res., vol. 13, pp. 281–305, 2012.
  39. 39. Ahmed E., Jones M., and Marks T. K., “An Improved Deep Learning Architecture for Person Re-Identification,” in Computer Vision and Pattern Recognition (CVPR), 2015, pp. 3908–3916.
  40. 40. Lee C. S., Baughman D. M., and Lee A. Y., “Deep Learning Is Effective for Classifying Normal versus Age-Related Macular Degeneration OCT Images,” Opthalmology Retin., vol. 1, no. 4, pp. 322–327, Jul. 2017, pmid:30693348
  41. 41. Asaoka R., Murata H., Iwase A., and Araie M., “Detecting Preperimetric Glaucoma with Standard Automated Perimetry Using a Deep Learning Classifier,” Ophthalmology, vol. 123, no. 9, pp. 1974–1980, Sep. 2016, pmid:27395766
  42. 42. Wang J. et al., “Joint retina segmentation and classification for early glaucoma diagnosis.,” Biomed. Opt. Express, vol. 10, no. 5, pp. 2639–2656, May 2019, pmid:31149385