Comparisonal study of Deep Learning approaches on Retinal OCT Image

Comparisonal study of Deep Learning approaches on Retinal OCT Image


In medical science, the use of computer science in disease detection and diagnosis is gaining popularity. Previously, the detection of disease used to take a significant amount of time and was less reliable. Machine learning (ML) techniques employed in recent biomedical researches are making revolutionary changes by gaining higher accuracy with more concise timing. At present, it is even possible to automatically detect diseases from the scanned images with the help of ML. In this research, we have taken such an attempt to detect retinal diseases from optical coherence tomography (OCT) X-ray images. Here, we propose a deep learning (DL) based approach in detecting retinal diseases from OCT images which can identify three conditions of the retina. Four different models used in this approach are compared with each other. On the test set, the detection accuracy is 98.00% for a vanilla convolutional neural network (CNN) model, 99.07% for Xception model, 97.00% for ResNet50 model, and 99.17% for MobileNetV2 model. The MobileNetV2 model acquires the highest accuracy, and the closest to the highest is the Xception model. The proposed approach has a potential impact on creating a tool for automatically detecting retinal diseases.

Disease Detection, Deep Learning, Image Analysis

* \AddToShipoutPictureBG*\AtPageUpperLeft

I Introduction

Retinal diseases are the most common cause of losing eye-sight at an early age. Most of them cause visual symptoms in the retina, a thin layer of tissue which is situated on the inside back wall of one’s eye. Some of the retinal diseases are- choroidal neovascularization (CNV), drusen, and diabetic macular edema (DME) which are shown in Fig. 2. These diseases can be detected from the retinal OCT scanned image. In Normal OCT scan, the choroid layer has no hollow space, no damage in the fovea, or extra material between membrane and pigment. Retinal diseases can affect the choroid layer causing an effect on eye-sight severely. The layer can be teared up or bloated in some places or hole can be created in some areas. Myopia, vision loss, macular degeneration, etc., can occur because of the affected layer. Hence, this article’s primary focus lies in the detection of retinal diseases from OCT scanned images.

In the case of traditional approaches, automatic detection of retinal diseases includes a preprocessing unit with image quantization, downsampling, and segmentation. After preprocessing, these images are then used to train a shallow neural network which takes a considerable amount of time as a huge amount of data is required [4]. Because of the limitations of the shallow network, nowadays transfer learning techniques (one of the DL techniques) have become more widespread [11], [9].

In this work, four CNN models- Vanilla CNN, MobileNetV2, ResNet50, and Xception network are employed to detect the category of diseases from retinal OCT scanned images. The images are mainly of 4 categories and have variable sizes. By preprocessing, each image is transformed into a definite shape. These transformed images are passed to the model for training. After training, these models are evaluated on a separated test data to measure performance. The performance metrics of each model are depicted in Fig. 4. No extra image processing based feature extraction methods is adopted in this work. For this reason, it can work with any data. The primary goal of this work is to help the patient and doctor and make the diagnosis automated and faster. Another goal is to enhance the performance of the analysis by increasing accuracy.

The specific contributions of our work are as follows:

  • We use images of variable sizes for training the models. To the best of our knowledge, for the first time such kind of work is done to handle variable sizes of OCT images to train the models.

  • We train the models using batches of data from different categories at the same time to maintain randomness (including around 44.57% CNV, 13.59% DME, 10.32% Drusen, and 31.52% Normal data among 84,452 images).

  • We demonstrate different robust models of deep convolutional neural network for detection of diseases and demonstrate a comparison of performance measure among them.

The remaining section consists of the following: Some related works are discussed in section II. In section III, the architecture for this work, description of the dataset, and evaluation criteria are given. Experimental analysis and results are briefly discussed in section IV. At last, the paper is concluded in section V.

Ii Related Works

In [15], Srinivasan et al. propose a fully automated way using sparsity-based block-matching and 3D-filtering (BM3D) to remove noise for detecting retinal diseases like- DME and age-related macular degeneration (AMD) from the spectral domain (SD) OCT volume images. In this method, the fraction of volume correctly classified for AMD and DME classes are 100%, which is the indication of overfitting problem. However, in the case of a regular class, it is 86.67%.

Burlina et al. describe a technique in which they used a pre-trained DCNN to detect AMD in [1]. This approach has achieved quite an excellent preliminary result.

By using B-scans of OCT, a fully automated method to detect lesion activity of AMD is studied by Chakravarthy et al. in [2]. They have tested retinal specialist (RS) grading versus Notal OCT analyzer (NOA) for faster treatment purpose and got an accuracy of 91%.

CNV, DME, Drusen, and their medical diagnoses are identified using CNNs and transfer learning by Kermany et al. in [8]. Collected images quality is reviewed initially and then is fed into the AI system. They have achieved accuracy of 96.6% in a multi-class comparison.

In [16], two-step framework to detect DME in OCT scans using patched based deep learning (DL) is proposed by Vahadane et al. They have used Dijkstra’s shortest path algorithm for detecting candidate patches for fluid-filled regions, Otsu thresholding technique for extraction. For hard regions, local peak detection is used. After these, they have used a DCNN to predict the label of those patches. The accuracy of this approach is better than using a frame-level DCNN classifier, which is proven by them.

Using artificial intelligence approach, fully automated way to detect and quantify of macular fluid - Schlegl et al. describe intraretinal cystoid fluid (IRC) and subretinal fluid (SRF) in macular OCT scan which causes AMD and DME in [14]. DCNN is used for mapping images to corresponding labels, and high accuracy is achieved in that process.

In [12], Rong et al. have used CNNs and suggested a classification method of surrogate-assisted to classify retinal OCT images automatically. They have reduced the noise from the image and applied thresholding and morphological dilation for extraction purpose. From there, they have generated surrogate images which are used for training. The accuracy is 97.83% for their local database, and 98.56% for the public database.

Li et al. have used the DL method VGG-16 network in [10] for detecting retinal disorders in a fully automated way. Image normalization is used to label the images, and image denoising is avoided to get more accuracy for original data. Using this process, they have achieved prediction accuracy up to 98.6%.

By using VGG-19 model, automated AMD detection combining OCT and fundus images are shown by Yoo et al. in [17] where five different models are used. An accuracy of 82.6% and 83.5% are achieved from only OCT image-based DL model and only fundus image based DL model whereas, in case of a combined image model, the accuracy is 90.5%.

Iii Methodology

The whole architecture for the analysis of OCT image data requires several phases which are described in the following. Fig. 1 illustrates the overall workflow of the research.

Fig. 1: Working Procedure

At the first stage of the workflow, raw image data goes through streaming and preprocessing pipeline. If a large amount of data are processed together for training, it causes overload on RAM, which may result in a system crash because of memory limitations. However, in this case, we have used data streaming, which takes data from the directory as batches. Our streamer can automatically detect the folders of classes for training and testing from the given directory for reading files. So, less code is required for reading a file than other cases. While streaming of the data from the directories, some filtering steps are executed dynamically at runtime. These filtering processes are data augmentation, data shuffling, and resizing of data. Augmentation step creates synthetic images by varying positioning, rotating, flipping, and shearing the real image data. After augmentation, we have resized both synthetic and original image into a fixed size to make every image uniform. On this uniform-sized images, we have applied the normalization process. At the final stage of data preprocessing pipeline, we have shuffled and partitioned the whole resulted images into training and testing sets. Shuffling is necessary for removing biases towards one particular class. The ratio of training, validation, and testing data split is 98.816:0.038:1.146. We have trained DL and transfer learning techniques vanilla CNN, MobileNetV2, ResNet50, and XCeption network with the training data in the training model phase of the workflow. After each epoch of training, we get the training accuracy curve of a model. At the same time, the validation accuracy curve of that model is also generated using validation dataset. At the last stage of the workflow, we have used different evaluation metrics and test data to evaluate the models that have been trained on the training data after 15 epochs of each model, which is known as the final model. Evaluation metrics consist of a confusion matrix, accuracy, precision, sensitivity, and f1 score for our proposed work.

Iii-a Dataset Collection and Description

We have collected the dataset from popular website Kaggle1. The whole dataset is organized into three folders (train, test, validation) containing sub-folders for four image categories (NORMAL, CNV, DME, DRUSEN) having total 84,484 X-Ray images (JPEG) with various shapes. Among these images for training and validation purpose, 83,484 and 32 images are used. We set aside 968 images for the test set. In Fig. 2, one figure from each category is given. From the figures, we can see the normal patient’s eye in Fig. 2(d) and rest of the figures are of the infected eye.

Fig. 2: (a) CNV (b) DME (c) DRUSEN (d) NORMAL

New blood vessels which are created in the choroid layer of the eye is known as Choroidal neovascularization (CNV). The blood vessels are created from a break in the Bruch membrane. It is the main cause of losing central vision. Central Vision is the process in which light changes into nerve signals using millions of cells that inform the brain what the person is seeing. CNV is associated with extreme myopia, malignant myopic degeneration, and age-related developments. The break in the membrane can be seen as a hollow space in the retinal OCT x-ray in Fig. 2(a).

Diabetic macular edema (DME) is a complication of diabetes. Fluid accumulation in the retina’s macula due to leaking blood vessels is the cause of DME. This disease makes the macula swell and affects the center of the macula, known as the fovea. In the OCT scanned image given in Fig. 2(b), the defect in fovea can be seen as small holes in the membrane. It causes blurred vision, wavy vision, double vision, a sudden change in eye floaters, and faulty perception of colors. It is also associated with low levels of protein in body fluids, high blood pressure, retention of fluid, and high levels of fats in the blood.

Between Bruch’s membrane and the retinal pigment epithelium of the eye, if tiny yellow or white accumulations of extracellular material are build up, it is known as Drusen. They are like tiny pebbles of debris. Drusen is associated with aging and macular degeneration which causes central vision loss.

The dataset distribution among these four categories is given in the Table I.

Data Training set Testing Set % of % of
Frequency Frequency Training Testing
CNV 37205 242 44.57 25
DME 11348 242 13.59 25
DRUSEN 8616 242 10.32 25
NORMAL 26315 242 31.52 25
Total 83484 968 100 100
TABLE I: Frequency Distribution of Retinal OCT Image

Iii-B Deep Learning Models

DL approaches are so far, the most stable strategy to handle image type data. Alongside the convolutional neural network, models like XCeption, MobileNetV2, ResNet have been employed in this work.

Convolutional Neural Network (CNN)

In the Convolutional Neural Network (CNN), each image data is flattened into vectors without losing features. As a result, the number of parameters involved and reusability of weights is reduced, and functional prediction result is obtained. The kernel, having the same depth as an image for each channel, is used to maintain valid and same padding and is shared by all the nodes of a layer. So, the resultant output has fewer nodes than a vanilla neural network. To reduce the spatial size, Pooling layer is also used, which can be of two types- max pooling and average pooling. Between these two pooling, max-pooling is better because it returns the maximum value from the part of the image covered by the kernel and performs as a noise suppressant. This algorithm is used in various architecture like- ResNet, Xception, VGGNet, and other DL techniques.

The vanilla CNN employed for this work developed with one input layer, eleven hidden layers, and one output layer. The model architecture for CNN is as following:

  • Input Layer (Shape = 150x150x3)

  • Convolutional Layer (Filter Size = 3x3, # of filters = 64, activation = Relu)

  • Max Pooling Layer (Filter size = 2x2)

  • Convolutional Layer (Filter Size = 3x3, # of filters = 64, activation = Relu)

  • Max Pooling Layer (Filter size = 2x2)

  • Convolutional Layer (Filter Size = 3x3, # of filters = 128, activation = Relu)

  • Max Pooling Layer (Filter size = 2x2)

  • Convolutional Layer (Filter Size = 3x3, # of filters = 128, activation = Relu)

  • Max Pooling Layer (Filter size = 2x2)

  • Flatten Layer

  • Dropout (Probability = .5)

  • Dense Layer (Shape = 512, activation = Relu)

  • Dense Layer (Shape = 4, activation = softmax)

Xception Net

The Xception stands for ’Extreme Inception’ which uses a modified depth wise separable convolution[3]. This modification is to have a pointwise convolution followed by a depth wise convolution instead of having the usual another way around. There are also residual connections added among the depth wise separable convolution layers. Finally, it takes inception like architecture with three flow networks added sequentially. The three flows are - entry flow, middle flow, and exit flow. In this work, following layers are embedded with the base architecture of Xception: Input Layer (Shape = 150x150x3), Flatten layer, Dense layer (Shape = 1024, activation = Relu), Dropout (Probability = .2), Dense layer (Shape = 4, activation = softmax).


In Residual network (ResNet), the layers are constructed as identity mapping, each taking the output of the previous layer as the input using shortcut connection. So, no extra parameter is needed. Between 18-layer and 34-layer net, 34-layer net has less error and better accuracy [7]. ResNet50 is Residual network with 50 layers where every two layers in the 34-layer net are replaced with 3-layer bottleneck block. For this work, the model architecture of ResNet50 is embedded with following layers: Input Layer (Shape = 150x150x3), Flatten layer, Dense layer (Shape = 1024, activation = Relu), Dropout (Probability = .2), Dense Layer (Shape = 4, activation = softmax).


MobileNetV2 uses inverted residual structure in which nonlinearities of thin layers are removed [13]. It works with two types of blocks. The basic building block is 19 bottleneck depth-separable convolution layers with residuals with stride 1 in one block and stride 2 in another block, which is preceded by fully convolutional layer with 32 filters. The first layer is 1x1 convolutional with Relu6, the second layer is depth wise convolution, and the third layer is 1x1 convolutional without nonlinearities using kernel size 3x3 for each layer. In the case of object detection, higher accuracy can be gained by this model. For this work, following layers are embedded with the MobileNetV2: Input layer (Shape = 150x150x3), Flatten layer, Dense layer (Shape = 1024, activation = Relu), Dropout (Probability = .2), Dense layer (Shape = 4, activation = softmax).

Iii-C Evaluation Criteria

Confusion Matrix

The Confusion matrix, also known as the error matrix, can be used for visualization of a model’s performance as the summary of prediction result for each class is accounted here [6]. We can easily find the confusion or errors made by a classifier by checking the confusion matrix table. From this table, we can get the value of True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN). Let us assume, Cl denotes a class out of four classes. The definition of TP, TN, FP, FN for Cl are as follows:

  • TP(Cl) = All the Cl instances that are predicted correctly as Cl.

  • TN(Cl) = All the instances of non Cl that are not predicted as Cl.

  • FP(Cl) = All the instances of non Cl that are predicted as Cl.

  • FN(Cl) = All the Cl instances that are not predicted as Cl.


The measurement of how accurately a classifier classifies a data is known as accuracy [5]. The equation for calculating accuracy is as follows:


Where, T.P. = True Positive, T.N. = True Negative, F.P. = False Positive and F.N. = False Negative.


Precision refers to the ratio of total correct positive results and total predicted positive results by the classifier. The equation of sensitivity is the following:



Sensitivity is known as the true positive rate or recall, which refers to the proportion of correct positive data points respect to all positive data points. The equation of sensitivity is the following:



In a model, the weighted average of Precision and Recall is known as F1 score [5]. The equation of the F1 score is the following:


Iv Experimental Analysis and Result Analysis

Iv-a Experimental Setup

Google Colab solely leverages this classification task. Colab consists of 1xTesla K80 (2496 CUDA cores), 1xsingle core hyper threaded Xeon Processors @2.3Ghz, 45MB Cache, 12.6 GB available RAM, and 320 GB available Disk Space.

Iv-B Result Analysis

Table II describes the confusion matrix of our four models after training and testing. From the confusion matrix table, we found that Xception and MobileNetV2 can predict more accurately than ResNet50 and CNN. By using Eq. 1, 2, 3, 4, we have calculated the accuracy, precision, sensitivity and F1 Score respectively for each classifiers. The values are given in the Table III. From Table III, it is found that the training accuracy in all models is 89% or above where the testing accuracy is 97% or above. The training accuracy is higher in Xception, which is 93.90%, and the testing accuracy is higher in MobileNetV2, which is 99.17%. The precision in training is between 0.92 to 0.95, and in the case of testing, it is between 0.97 to 0.99, including the boundary values. For both cases, Xception has the highest precision, and MobilenetV2 has the highest precision only in testing. The Sensitivity and F1 score in training are from 0.92 to 0.95, and in testing is from 0.97 to 0.99. Sensitivity and F1 score of 0.95 is achieved in Xception for training, and 0.99 is achieved for testing in Xception and MobileNetV2. From the four models, Xception and MobileNetV2 have yielded the best results.




CNV 35638 468 807 292
DME 527 9644 30 1147
DRU 988 50 6229 1349
NRL 109 219 48 25939


CNV 240 2 0 0
DME 4 233 0 5
DRU 0 0 242 0
NRL 0 0 5 237



CNV 35467 197 1391 150
DME 288 9813 59 1188
DRU 272 11 7651 682
NRL 36 25 145 26109


CNV 241 1 0 0
DME 3 238 0 1
DRU 4 0 238 0
NRL 0 0 0 242



CNV 36093 251 776 85
DME 806 9184 43 1315
DRU 1323 24 6244 1025
NRL 354 99 209 25653


CNV 241 1 0 0
DME 6 235 0 1
DRU 18 0 222 2
NRL 1 1 0 240



CNV 35399 182 1367 257
DME 427 10205 27 689
DRU 389 27 7039 1161
NRL 31 152 61 26071


CNV 241 1 0 0
DME 2 240 0 0
DRU 3 0 238 1
NRL 0 0 1 241
TABLE II: Confusion Matrix (Training and Testing)
Evaluation Classifiers
Metrics CNN Xception ResNet-50 MobileNet-V-2


Accuracy 0.90 0.9390 0.89 0.9388
Precision 0.93 0.95 0.92 0.94
Sensitivity 0.93 0.95 0.92 0.94
F1 Score 0.93 0.95 0.92 0.94


Accuracy 0.98 0.9907 0.97 0.9917
Precision 0.98 0.99 0.97 0.99
Sensitivity 0.98 0.99 0.97 0.99
F1 Score 0.98 0.99 0.97 0.99
TABLE III: Evaluation metrics of our study

For this work, we have used 15 epochs for training and validation purposes. The training accuracy vs. epochs graph is shown in Fig. 3. In the case of CNN, after first epoch the curve has decreased until the fourth epoch; then in the fourth epoch, it has increased and then decreased again. The Xception’s curve has increased till the second epoch, then kept on decreasing and increasing until eighth epoch. After eighth epoch, it has become saturated. The MobileNetV2’s curve has increased until the 2nd epoch. Then it has become saturated with higher accuracy like Xception. The Resnet50’s curve has shown encouraging result until the fourth epoch as the accuracy has abruptly risen. Then the curve’s changing has become slower and has reached around the saturation in sixth epoch with lower accuracy than Xception and MobileNetV2. Thus, the Xception and MobileNetV2 are more promising in case of training than the other two because of their higher accuracy.

Fig. 3: Training accuracy vs Epochs

The graph of validation accuracy for 15 epochs is shown in Fig. 4. The curve of vanilla CNN is upward and downward, and after a time, it has reached its saturation. The Xception’s curve has reached the highest accuracy in most of the epochs. The MobileNetV2’s curve has been changing abruptly upwards and downwards, and in the last epoch, it reached close to the Xception’s accuracy. In the case of ResNet50’s curve, drastic change of accuracy can be seen till the 4th epoch. After that, the accuracy has increased and decreased steadily, and in the last epoch, it has reached near CNN’s accuracy. So, from Fig. 4, it can be said that Xception and MobileNetV2 have higher accuracy than the other two in testing.

Fig. 4: Validation accuracies vs Epochs

From the Table IV, we can find that Kermany et al. has gained 96.60% prediction accuracy with a sensitivity of 97.80% in detection of retinal diseases using convolutional neural network [8], where Li et al. has achieved a prediction accuracy of 98.60% having sensitivity of 97.80% for predicting retinal diseases in a fully automated way [10]. In our method, to detect CNV, DME, and Drusen- these three diseases using the best model, accuracy of 99.17% with a sensitivity of 99.00% is achieved. These three works are on the same dataset. So, from the quantitative result, it is clear that our method is better among all the methods given in Table IV as it has the highest accuracy.

Method Accuracy Sensitivity Precision F1-score
Our method 99.17% 99.00% 99.00% 0.99
Kermany et al. [8] 96.60% 97.80% N/A N/A
Li et al. [10] 98.60% 97.80% N/A N/A
TABLE IV: Proposed Comparative Analysis with Some Contemporary Methods

V Conclusions

In this experimental study, we demonstrate a better way to detect some retinal diseases from OCT X-ray and also show the comparison with other works [8], [10]. Our primary contribution is to work with variable sized images. We explored multiple deep learning models [3], [7], [13] to present the performances in terms of accuracy, sensitivity, precision, and f1 score [6], [5]. We can claim that our approach can detect the discussed retinal diseases better as the accuracy is higher than state of the art. There is a limitation regarding the distribution of classes; that is, all classes are not equally distributed. If it can be overcome, the performance will increase. We have not used other image processing techniques. Maybe the use of image processing techniques could have led to an exciting result, which can make our work better and increase performance. We are leaving these issues for future works.




  1. P. Burlina, D. E. Freund, N. Joshi, Y. Wolfson and N. M. Bressler (2016) Detection of age-related macular degeneration via deep learning. In 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI), pp. 184–188. Cited by: §II.
  2. U. Chakravarthy, D. Goldenberg, G. Young, M. Havilio, O. Rafaeli, G. Benyamini and A. Loewenstein (2016) Automated identification of lesion activity in neovascular age-related macular degeneration. Ophthalmology 123 (8), pp. 1731–1736. Cited by: §II.
  3. F. Chollet (2017) Xception: deep learning with depthwise separable convolutions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1251–1258. Cited by: §III-B2, §V.
  4. M. Goldbaum, S. Moezzi, A. Taylor, S. Chatterjee, J. Boyd, E. Hunter and R. Jain (1996) Automated diagnosis and image understanding with object extraction, object classification, and inferencing in retinal images. In Proceedings of 3rd IEEE International Conference on Image Processing, Vol. 3, pp. 695–698. Cited by: §I.
  5. M. Hasan, I. Islam and K. A. Hasan (2019) Sentiment analysis using out of core learning. In 2019 International Conference on Electrical, Computer and Communication Engineering (ECCE), pp. 1–6. Cited by: §III-C2, §III-C5, §V.
  6. M. Hasan, M. M. Islam, I. Islam and M. Hashem (2019) Attack and anomaly detection in iot sensors in iot sites using machine learning approaches. Internet of Things, pp. 100059. Cited by: §III-C1, §V.
  7. K. He, X. Zhang, S. Ren and J. Sun (2016) Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778. Cited by: §III-B3, §V.
  8. D. S. Kermany, M. Goldbaum, W. Cai, C. C. Valentim, H. Liang, S. L. Baxter, A. McKeown, G. Yang, X. Wu and F. Yan (2018) Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 172 (5), pp. 1122–1131. Cited by: §II, §IV-B, TABLE IV, §V.
  9. A. Krizhevsky, I. Sutskever and G. E. Hinton (2012) Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pp. 1097–1105. Cited by: §I.
  10. F. Li, H. Chen, Z. Liu, X. Zhang and Z. Wu (2019) Fully automated detection of retinal disorders by image-based deep learning. Graefe’s Archive for Clinical and Experimental Ophthalmology 257 (3), pp. 495–505. Cited by: §II, §IV-B, TABLE IV, §V.
  11. Y. Liu, M. Chen, H. Ishikawa, G. Wollstein, J. S. Schuman and J. M. Rehg (2011) Automated macular pathology diagnosis in retinal oct images using multi-scale spatial pyramid and local binary patterns in texture and shape encoding. Medical image analysis 15 (5), pp. 748–759. Cited by: §I.
  12. Y. Rong, D. Xiang, W. Zhu, K. Yu, F. Shi, Z. Fan and X. Chen (2019) Surrogate-assisted retinal oct image classification based on convolutional neural networks. IEEE journal of biomedical and health informatics 23 (1), pp. 253–263. Cited by: §II.
  13. M. Sandler, A. Howard, M. Zhu, A. Zhmoginov and L. Chen (2018) Mobilenetv2: inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510–4520. Cited by: §III-B4, §V.
  14. T. Schlegl, S. M. Waldstein, H. Bogunovic, F. Endstraßer, A. Sadeghipour, A. Philip, D. Podkowinski, B. S. Gerendas, G. Langs and U. Schmidt-Erfurth (2018) Fully automated detection and quantification of macular fluid in oct using deep learning. Ophthalmology 125 (4), pp. 549–558. Cited by: §II.
  15. P. P. Srinivasan, L. A. Kim, P. S. Mettu, S. W. Cousins, G. M. Comer, J. A. Izatt and S. Farsiu (2014) Fully automated detection of diabetic macular edema and dry age-related macular degeneration from optical coherence tomography images. Biomedical optics express 5 (10), pp. 3568–3577. Cited by: §II.
  16. A. Vahadane, A. Joshi, K. Madan and T. R. Dastidar (2018) Detection of diabetic macular edema in optical coherence tomography scans using patch based deep learning. In 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pp. 1427–1430. Cited by: §II.
  17. T. K. Yoo, J. Y. Choi, J. G. Seo, B. Ramasubramanian, S. Selvaperumal and D. W. Kim (2019) The possibility of the combination of oct and fundus images for improving the diagnostic accuracy of deep learning for age-related macular degeneration: a preliminary experiment. Medical & Biological Engineering & Computing 57 (3), pp. 677–687. Cited by: §II.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
Add comment
Loading ...
This is a comment super asjknd jkasnjk adsnkj
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test description