Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Atrous residual convolutional neural network based on U-Net for retinal vessel segmentation

  • Jin Wu,

    Roles Methodology, Resources, Software, Supervision, Writing – original draft

    Affiliation School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan, China

  • Yong Liu ,

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Methodology, Software, Validation, Visualization, Writing – original draft, Writing – review & editing

    141001@xxmu.edu.cn

    Affiliations School of Information Science and Engineering, Wuhan University of Science and Technology, Wuhan, China, School of Medical Engineering, Xinxiang Medical University, Xinxiang, China

  • Yuanpei Zhu,

    Roles Project administration, Resources, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation School of Physics and Electronic Engineering, Xinxiang University, Xinxiang, China

  • Zun Li

    Roles Supervision, Writing – original draft, Writing – review & editing

    Affiliation School of Physics and Electronic Engineering, Xinxiang University, Xinxiang, China

Abstract

Extracting features of retinal vessels from fundus images plays an essential role in computer-aided diagnosis of diseases, such as diabetes, hypertension, and cerebrovascular diseases. Although a number of deep learning-based methods have been used in this field, the accuracy of retinal vessel segmentation remains challenging due to limited densely annotated data, inter-vessel differences, and structured prediction problems, especially in areas of small blood vessels and the optic disk. In this paper, we propose an ARN model with a atrous block to address these issues, which can avoid the loss of data structure, and enlarge the receptive field, so that each convolution output contains a larger range of information. In addition, we also introduce residual convolution network to increase the network depth and improve the network performance.Some key parameters are used to measure the feasibility of the model, such as sensitivity (Se), specificity (Sp), F1-score (F1), accuracy (Acc), and area under each curve (AUC). Experimental results on two benchmark datasets demonstrate the effectiveness of the proposed methods, which accuracy are 0.9686 on the DRIVE and 0.9746 on the CHASE DB1. The segmentation structure can assist the doctor in diagnosis more effectively.

Introduction

Diseases such as diabetes, hypertension, and diseases of the retina are shown in retinal vascular images [1]. The analysis of the number, angle, branch, and curvature of retinal blood vessels can provide a valuable basis for clinical diagnosis [2], for the purpose of early prevention, diagnosis, and treatment. With the development and popularity of optical coherence tomography (OCT) imaging technology and the increasing emphasis on early disease diagnosis, the number of fundus images is increasing rapidly, and their analysis will require much time and effort [3]. In addition, differences in image acquisition procedures between machines and institutions may lead to huge differences in resolution, noise, and tissue appearance, which increase the difficulty of analysis [4].

To meet the need of this work, a fast and automatic segmentation method for retinal vascular images came into being [5,6]. It can improve the cutting efficiency and accuracy, reduce the waiting time of patients and save medical resources [7,8]. It can also provide basis for fundus image registration, arteriovenous classification and biometric recognition [9]. Notably, deep learning methods have performed better than traditional methods [1013].

Among the unsupervised methods, filtering, morphological transformation and model-based algorithms are dominant [14]. The method was evaluated on DRIVE and STARE databases and returned accuracies of 0.945 and 0.9486 respectively.Wavelet transform was used by Akram et al. and Soares et al. in retinal vascular segmentation [15], and they achieved 94.4% accuracy in STARE. 2D Gabor Wavelet and Gaussian mixture models are used in their approach. In particular, this approach is heavily influenced by the quality of the image. The entropy of some particular antennas with a pre-fractal shape, harmonic sierpinski gasket and weierstrass-mandelbrot fractal function were studied, and the result indicated that their entropy is linked with the fractal geometrical shape and physical performance [16,17], and they achieved 94.3% and 94.4% accuracy in DRIVE and STARE. This method is a kind of unsupervised technique, and the calculation is fast, but the accuracy is limited. Frangi et al. proposed a multi-scale enhanced-vessel filtering method to enhance vascular and vascular-like patterns, in which second-order local structural features were used [18]. Sato et al. applied three-dimensional (3D) multi-scale line filtering to the segmentation of cerebrovascular, bronchial and liver vessels [19]. This method can improve the continuity of the circuit structure and reduce the noise, but the calculation is large and slow. Jiang et al. proposed a universal vessel segmentation framework based on adaptive local threshold and applied it to retinal vessel segmentation [20], they observed 65% true positive rate, and This method has more parameters and slower operation. Zhang et al. proposed a filter based segmentation method for retinal vessels, which uses a locally adaptive derivative filter [21], and they achieved 94.76% and 95.54% accuracy in DRIVE and STARE. Azzopardi et al. improved COSFIRE operator detection and applied it in retinal segmentation [22], and they achieved 94.27% and 94.11% accuracy in DRIVE and CHASE_DB1. Zhao et al. designed a new retinal vessel segmentation model using an infinite parameter active contour model with mixed regional information [23], and they achieved 95.4% and 95.6% accuracy in DRIVE and STARE. Liang et al. proposed a level set method for vessel segmentation based on regional energy fitting information and shape prior probability [24] and they achieved 95.03% and 95.36% accuracy in DRIVE and STARE. The framework of unsupervised segmentation method always uses the filtering method which is sensitive to blood vessels or vessel-like, which will lead to the incomplete blood vessels and the misidentification of vessel-like parts. Moreover, the parameter setting has great influence on the final segmentation result.

For the supervised methods, ground truth must be used to train the classifier, and then the classifier can be used to extract the blood vessels. The features of retinal blood vessels can be extracted by multiple methods [25,26]. Traditional machine learning methods for training classifiers use k nearest neighbors, adaboost, random forest and other methods [27], and they achieved 92.9% accuracy in DRIVE. Orlando et al. proposed a fully connected conditional random field model for retinal vascular segmentation, using structured output support vector machine learning model parameters as an example to improve the effect [28], and they achieved 0.8741 and 0.8628 G-mean value in DRIVE and STARE. Zhang et al. applied retinal vascular segmentation through filtering and wavelet transform strategy, and used random forest training strategy [29], and they achieved 94.66% and 95.47% accuracy in DRIVE and STARE. In the above methods, the key feature selection has a great influence on the final segmentation result, such as whether features are independent or easy to identify. However, these features must be selected through people’s experience, The features need to be selected manually according to the experiment; so there is still much work to be done to perfect their shortcomings.

Benefiting from the rapid development of computer hardware, convolutional neural networks (CNNs) [30] have become the main machine learning method. Many CNN-based classification and detection methods have been proposed, which have facilitated the rapid development of medical assisted diagnosis methods. In the process of retinal vascular segmentation, the proportion of vascular areas, especially capillaries, is relatively small. To achieve a better segmentation effect, it is necessary to increase the number of training sets and the amount of training time. However, the number of training sets is limited in existing public datasets. To solve this problem, Ronneberger et al. proposed U-Net [31], which combines coarse and fine features through skip connections, can achieve better accuracy(95.34% and 95.78% accuracy in DRIVE and STARE) with fewer training sets. Many methods based on U-Net have achieved good results, but there are still problems such as low accuracy, poor sensitivity, and segmentation area error, especially the loss of vessel branch points, intersection points, and small vessels.The results by other authors are summarized in Table 1.

We propose ARU-Net, a deep learning model to automatically segment retinal blood vessels in fundus images. The model leverages the strengths of U-Net, cascaded atrous convolution, and residual blocks enriched with squeeze and excitation. Residual blocks [32] are used as building units to simplify the training process and help extract coarse and fine features from source images. Squeeze and excitation units are added to each remaining block for channel attention, adaptive feature recalibration, and increased feature power representation. The addition of a dilated convolution module can ensure global and multi-scale extraction. We evaluated our model on the publicly available DRIVE [33] and CHASE DB1 [34] datasets, and the results show that it is effective, and the performance is improved.

The proposed approach has the following contributions.

  1. We propose a U-Net model integrating modified residual blocks to improve network performance.
  2. An improved hybrid atrous convolution is used to increase the receptive field.

The rest of this paper is organized as follows: Section 2 analyze relevant literature. Section 3 presents the proposed method; Section 4 analyzes and discusses the experiment result; Section 5 concludes this study.

Related work

Due to the excellent performance of deep learning framework in retinal vascular segmentation, We will analyze ralated works using typical deep learning architectures.

U-Net

U-Net can be divided into three parts: left (down-sampling), middle (copy and crop), and right (up-sampling). The first part reduces the size of the picture through four down-sampling operations, which extract features from shallow information. The copy and crop part includes four splicing operations. This operation fuses characteristic deep and shallow information. In the up-sampling part, the picture is larger, and deep information is extracted through four up-sampling operations. In the process of up-sampling, the number of channels in the image is halved, which is contrary to the change of the number of channels in feature extraction in the left part [35]. The up-sampling process fuses the shallow information on the left and splices the features. A skip connection is used in U-Net at the same stage, ensuring that the recovered feature graph integrates more low-level features and features of different scales. In this way, multi-scale prediction and deep supervision can be carried out, and information such as edge recovery of segmentation maps can be more refined [36].

ResNet

It has been found that deeper network layers and smaller receptive fields can improve neural network performance. However, as the network structure deepens, two problems arise. First, vanishing and exploding gradients affect the convergence of training. Second is degradation. An increasing number of layers causes the model accuracy to decrease (which is not caused by overfitting), and the training and testing error both increase.To overcome these problems, the The residual network proposed by He et al. [37] shows significantly improved training characteristics, allowing previously unachievable network depths.

In contrast to the traditional convolutional or fully connected layer, ResNet has many bypass branches that connect the input directly to the following layer, so as to directly learn the residuals. This structure is also known as a shortcut connection. Such a structure can directly detour the input information to the output to protect its integrity. The network only needs to learn the input and output differences of that part, simplifying the learning objectives and decreasing difficulty.

Method

Using U-Net as the basic framework in medical image segmentation can solve the problem of small samples that commonly exist in such images. However, U-Net is composed of a contraction path that gradually reduces the spatial dimension of the image through down-sampling, and an expansion path that gradually restores the details and spatial dimension of the object through up-sampling. Therefore, after convolution and down-sampling, there will be gradient disappearance, structural information loss, and other problems.

In this study, We integrate a residual network (ResNet) and atrous convolution modules into the U-Net network in a new network structure, the atrous residual U-Net (ARU-Net), which can further expand the receptive field and improve the correlation between objects without losing information, thus improving the performance of vascular segmentation.This framework is shown in Fig 1. It consists of two phases of training and testing. In the training stage, color fundus images were preprocessed with grayscale transformation and normalization, and then used as training data. The network adjusts model weight parameters by iterative learning. Then save the weights. In the testing stage, the network reloads the saved weight information and makes predictions for the preprocessed data.

Modified residual block

To further improve the performance of the network, we include a squeeze-and-excitation block in ResNet [38], the difference from the original residual network is shown in Fig 2B and 2C. Squeeze (red box, Fig 2(C)) can change the spatial dimension of each input feature map from H×W squeeze to 1×1. We use global average pooling to achieve this. In the squeeze operation, we perform feature compression along the spatial dimension, turning each two-dimensional feature channel into a real number that has a global receptive field, and the dimension of the output matches the number of feature channels of the input. It represents the global distribution of responses on the characteristic channel and enables the layer close to the input to obtain the global receptive field, which is useful in many tasks.

(1)
thumbnail
Fig 2.

(a) Proposed ARU–Net Block, (b) standard residual block, (c) modified residual block.

https://doi.org/10.1371/journal.pone.0273318.g002

Where zc represents the channel descriptor for channel c, Fsq represents global average pooling, uc represents channel c of the input, and H and W represent the height and width of the input.

In excitation (green box, Fig 2(C)), the feature dimension is reduced to 1/r of the input, and raised back to the original dimension through a fully connected layer after ReLU activation. This method, which is more nonlinear and can better fit the complex correlation between channels than the method that directly uses a fully connected layer, greatly reduces the number of parameters and the amount of calculation.

Hybrid atrous convolution block

As is well known, atrous convolution can enlarge the receptive field [39,40]. When the convolution kernel is 3×3, 1-dilated and 2-dilated together can achieve the effect of a 7×7 convolution kernel. Similarly, when 4-dilated conv is followed by 1-dilated and 2-dilated conv, the receptive field can achieve the effect of a 15×15 convolution kernel. Compared with traditional convolution operations, the receptive field of atrous convolution grows exponentially.

However, when we only stack convolution with the same void rate many times, the kernel is not continuous, i.e., not all pixels are used for calculation. Therefore, the information is regarded as checkerboard, which will lose continuity, and does not work well for small objects.

We use a hybrid atrous convolution block to solve this problem, (2) where ri is the void rate of the i-th layer, and Mi is the maximum void rate at layer i. Assuming there are n layers, the default is Mn = rn. If we apply a k×k convolution kernel and our goal is that M2K, then we can cover all the holes using standard convolution.

Proposed U-net Block is shown in the Fig 2A. We replaced the original COV3×3 block in the original U-net with AR Conv Unit. The red arrow represents AR Conv Unit, it is also an important difference from U-Net, and we can see the specific algorithm flow from Algorithm 1.

Algorithm 1: Algorithm of the proposed ARU convolution unit

    Input: Feature map X

    Output: Feature map Y

    1: X1 = HAC(X) // Hybrid atrous convolution block with r[1,2,3]

    5: X2 = h(X)+F(x,w);xl+1 = f(x2)

    6: X3 = Ftr(X2) // Ftr: X→U, XRW’×H’×C, URW×H×C

    7: X4 = FSq(X3) // Shrinking feature maps∈RW×H×C through saptical dimentions

    8: X5 = Fex(X4,W) //Learnning WRC×C to explicitly model channel-association

    9: X6 = Fscale(X2,X5) //Reweighting the feature maps∈RW×H×C

    10: Y = concatenate(X1,X6)

Experiments

Implementation details

We evaluated the proposed ARU-Net on the DRIVE and CHASE DB1 retinal image datasets. DRIVE includes 40 color fundus images, with 20 for training and 20 for testing. CHASE DB1 contains 28 retinal fundus images. Although there is no initial division of training and testing sets, the first 20 are usually used for training, and the remaining eight for testing. Additionally, in Chase DB1, there are two sets of Ground-truth (GT) images, we chose the first group because they have more complete vessel details than broken ones. Images in DRIVE and CHASE DB1 have resolution of 565×584 and 999 × 960, respectively. To fit our network, we resize each image in DRIVE and CHASE DB1 to 592 × 592 and 1008 × 1008 by padding it with zero in four margins, but in evaluation, we crop the segmentation results to the initial resolution. Manually segmented binary vessel maps of both datasets provided by human experts can be applied as the ground truth.

The experiments have been conducted in a desktop computer with intel core i5-9400 processor CPU, 32GB RAM, and NVIDIA 1080Ti, 11 GB GPU. Adam optimization method was used to optimize the parameters. Since the phased training can speed up the convergence of the network, we adopted different parameters in the training process [41]. When using the Drive data set, we find that the first 200 epochs use learning rate of 0.003, followed by the learning rate of 0.0001 can achieve good results. And the same is true for CHASE DB1. The difference is that we set the batch size to be 2 and 1 respectively. The relevant model dimensions are listed in Table 2.

Evaluation metrics

To quantitatively evaluate our model, we compared the segmentation results with the corresponding ground truth; divided the results of each pixel into true positive (TP), false positive (FP), false negative (FN), and true negative (TN); and adopted sensitivity (Se), specificity (Sp), F1-score (F1), and accuracy (ACC) to evaluate the performance of the model.

(3)(4)(5)(6)(7)(8)

We also utilize the area under the ROC curve (AUC), where a value of 1 indicates perfect segmentation.

Results

The results in Table 3 are our results on public datasets, the DRIVE and the CHASE DB1. We compare the single use of residual networks, the single use of empty convolution and our approach. When the residual convolution block or the hybrid atrous convolution block is added separately, the segmentation performance of the network can be improved. The experimental results show that when the above two kinds of convolution are added to the network, the segmentation network can get better results.

Fig 3 shows some examples from our experiments. From the segmentation results of the two datasets, our results are very close to the gold standard, especially for some small blood vessels. We also verify the influence of different blocks on the results. According to the results in Table 1, our method is superior to the previous best in some key parameters. The highest ACC (0.9686%/0.9746%), the highest AUC(0.9842/0.9869), and the highest Se(0.8149/0.8420). That means residual convolution blocks and hybrid atrous convolution block are very useful for networks.

thumbnail
Fig 3. Some examples from two experiments on the DRIVE and CHASE DB1 datasets.

The first three columns are from the DRIVE dataset. The last three columns are from the CHASE DB1 datasets. (a) and (d) Color fundus image, (b) and (e) ground truth, (c) and (f) segmentations.

https://doi.org/10.1371/journal.pone.0273318.g003

The proposed method was also compared with U-Net in a segmentation experiment, and some examples of the results are shown in Figs 4 and 5. Fig 4 shows examples on the DRIVE dataset; Fig 4(A)–4(D) are the color fundus image, ground truth (manual annotation data), results by the proposed method, and results by U-Net. Fig 5 shows examples of the results on CHASE DB1. We can conclude from these results that our method can segment more vascular details, especially in capillary vessels (marked by a red box).

thumbnail
Fig 4.

Comparison of ARU–Net and U–Net on DRIVE dataset: (a) color fundus images; (b) ground truth; (c) segmentation result by ARU–Net; (d) segmentation result by U–Net.

https://doi.org/10.1371/journal.pone.0273318.g004

thumbnail
Fig 5.

Comparison of ARU–Net and U–Net on CHASE DB1 dataset: (a) color fundus images; (b) ground truth; (c) segmentation result by ARU–Net; (d) segmentation result by U–Net.

https://doi.org/10.1371/journal.pone.0273318.g005

Finally, we compared ARU-Net with several state-of-the-art methods on the DRIVE and CHASE DB1 datasets, with results as shown in Table 4, which shows that ARU-Net performs best on both datasets as measured by Se, Sp, ACC, F1, and AUC. In details, on the DRIVE and CHASE DB1, our model has the highest AUC (0.21%/0.09% higher than the best before), the highest accuracy (1.08%/0.85% higher than the best before) and the highest sensitivity. F1 and specificity are generally comparable. Hence, our method achieves state-of-the-art performance for retinal vessel segmentation.

thumbnail
Table 4. Results of different methods on DRIVE and CHASE DB1.

https://doi.org/10.1371/journal.pone.0273318.t004

In addition, adding multi-scale strategy can improve network performance, just like on the Drive dataset. The Se, Acc and AUC were improved (Se0.7772, ACC0.9553, AUC0.9759 respectively) [55,56]. In their study, an improved cross entropy loss function is applied, and uses CRFs as a post-processing strategy. The challenge is how to exploit the relationships between images at different scales. The segmentation result is greatly affected by the weight coefficient, which needs to be set manually. And compared with some other methods, the performance improvement is not obvious. For convolution neural network method with reinforcement sample learning strategy proposed by Guo et al. [57], Sp and ACC value was the lowest, and the final segmentation result was the worst. For U-net based on patch-based learning strategy, Se, Sp, ACC and AUC value were not the highest; however, the segmentation result was the best in the comprehensive evaluation.

Discussion and conclusion

In medical image segmentation, common methods of data augmentation include random slice, rotation and mirror image, etc [58,59]. In general, the accuracy of the model on the training set is significantly increased, but in the validation set, the accuracy is not significantly improved, that is to say, the generalization ability of the model is not substantially improved. In order to improve the overall performance of the network, such as Se, Sp, ACC, AUC, and AUC, it is necessary to adjust the structure of the network, change the supervision function and the optimizer.

Unlike other data augmentation approaches, our method operates on the entire image. This has proven to be beneficial as our model is faster than the above methods. In addition, our model is able to obtain more natural and continuous segmentation masks and capture more detailed features. Furthermore, the introduction of hybrid atrous convolution blocks and modified residual blocks in our framework made it possible to utilize image at multiple scales, with corresponding oversight at each scale, helping our model to efficiently aggregate the outputs of different stages.

We presented ARU-Net, a segmentation structure to which atrous and residual convolution were added. This unit enlarges the receptive field without losing resolution. We replaced ReLU with LeakyReLU in the downsampling process. We evaluated the method on the DRIVE and CHASE DB1 benchmark datasets, and accuracy and sensitivity metrics demonstrated that our model can segment fundus vessels better than other models. The branches of many vessels, including very small vessels, were correctly segmented. Fundus diseases often reflect changes in the small shape of blood vessels.

Therefore, the above methods can help doctors to diagnose diseases. However, the segmentation rupture of blood vessels can still occur in images with lesions. In the future, we’ll continue to explore how to ameliorate the problem of broken blood vessels, so that the segmentation results are closer to the real.The experimental results on the DRIVE and CHASE DB1 datasets are shown in Figs 6 and 7.

References

  1. 1. Sohini R.; Dara D.K.; Keshab K.P. Blood Vessel Segmentation of Fundus Images by Major Vessel Extraction and Sub-Image Classification. IEEE J. Biomed. Health Inform. 2015, 19, 1118–1128.
  2. 2. Fraz Muhammad Moazam, et al. Blood vessel segmentation methodologies in retinal images-asurvey. Computer methods and programs in biomedicine 108.1 (2012): 407–433. pmid:22525589
  3. 3. Orujov F., et al. Fuzzy based image edge detection algorithm for blood vessel detection in retinal images. Applied Soft Computing (2020).
  4. 4. S Kumar A Adarsh, et al. An automated early diabetic retinopathy detection through improved blood vessel and optic disc segmentation. Optics & Laser Technology 121(2020):105815–105815.
  5. 5. Mondal SS, Mandal N, et al. Blood vessel detection from Retinal fundas images using GIFKCN classifier. Procedia Computer Science 167(2020):2060–2069.
  6. 6. Kushol R., Salekin M. S. Rbvs-Net: A Robust Convolutional Neural Network For Retinal Blood Vessel Segmentation. 2020 IEEE International Conference on Image Processing (ICIP) IEEE, 2020.
  7. 7. Solomon S. D., Chew E., Duh E. J., Sobrin L., Sun J. K.,VanderBeek B. L., et al., Diabetic retinopathy: a position statement by the american diabetes association. Diabetes. Care. 2017.40(3):412–418. pmid:28223445
  8. 8. Aghamohamadian-Sharbaf M., Pourreza H. R., and Banaee T., A novel curvature-based algorithm for automatic grading of retinal blood vessel tortuosity. J. Biomed. Health. Inform. 2016.20(2):586–595. pmid:25622332
  9. 9. Sohini R., Dara D.K., Keshab K.P. Blood Vessel Segmentation of Fundus Images by Major Vessel Extractionand Sub-Image Classification. IEEE J. Biomed. Health Inform. 2015, 19, 1118–1128. pmid:25014980
  10. 10. Guariglia E. Entropy and Fractal Antennas. Entropy 2016, Vol 18, Pages 84.
  11. 11. Sahlsten J., et al. Deep Learning Fundus Image Analysis for Diabetic Retinopathy and Macular Edema Grading. Scientific Reports (2019). pmid:31341220
  12. 12. Sakaguchi, A., Renjie, W., and Kamata, S., Fundus image classification for diabetic retinopathy using disease severity grading. In: Proceedings of the 2019 9th International Conference on Biomedical Engineering and Technology, pp(2019)190–196.
  13. 13. Shen Y., Sheng B., Fang R., Li H., Dai L., Stolte S., et al., Domain-invariant interpretable fundus image quality assessment, Vol. 61, p. 101654, 2020.
  14. 14. Adapa D., et al. A supervised blood vessel segmentation technique for digital Fundus images using Zernike Moment based features. PLOS ONE 15.3(2020):e0229831. pmid:32142540
  15. 15. Akram M. U., Atzaz A., Aneeque S. F., and Khan S. A., Blood vessel enhancement and segmentation using wavelet transform, in Proc. International Conf. on Digital Image Processing, 2009, pp. 34–38.
  16. 16. Jiang X, Mojon D. Adaptive local thresholding by verification-based multi-threshold probing with application to vessel detection in retinal images. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003, 25(1):131–137.
  17. 17. Fraz M. M., et al. An approach to localize the retinal blood vessels using bit planes and centerline detection. Computer Methods & Programs in Biomedicine 108.2(2012):600–616. pmid:21963241
  18. 18. Frangi A.F.; Niessen W.J.; Vincken K.I.; Viergever M.A. Multiscale vessel enhancement filtering. In Proceedings of the Imitational Conference on Medical Image Computing and Computer Assisted Intervention, Cambridge, MA, USA, 11–13 October 1998; Springer: Berlin, Germany, 1998; pp. 130–137. https://doi.org/10.1007/bfb0056195
  19. 19. Sato Y.; Nakaiima S.; Atsumi H.; Koller T.; Gerig G.; et al. 3D Multi-scale line filter for segmentation and visualization of curvilinear structures in medical images. Med. Image Anal. 1998, 2, 143–168. pmid:10646760
  20. 20. Jiang X., and Mojon D. Adaptive local thresholding by verification-based multithreshold probing with application to vessel detection in retinal images. IEEE Transactions on Pattern Analysis and Machine Intelligence 25.1(2003):131–137.
  21. 21. Zhang J., et al. Robust retinal vessel segmentation via locally adaptive derivative frames in orientation scores. IEEE Transactions on Medical Imaging 35.12(2016):1–1. pmid:27514039
  22. 22. Azzopardi G., et al. Trainable COSFIRE filters for vessel delineation with application to retinal images. Medical Image Analysis 19.1(2015):46–57. pmid:25240643
  23. 23. Zhao Y., et al. Automated Vessel Segmentation Using Infinite Perimeter Active Contour Model with Hybrid Region Information with Application to Retinal Images. IEEE Transactions on Medical Imaging 34.9(2015):1797–1807. pmid:25769147
  24. 24. Li-Ming Liang, et al. Retinal Vessel Segmentation Using Level Set Combined with Shape Priori. Chinese Journal of Computers (2018), 41, 1678–1692.
  25. 25. Guido R. C. Practical and Useful Tips on Discrete Wavelet Transforms. IEEE Signal Processing Magazine 32.3(2015):162–166.
  26. 26. Guido R., Addison P., and Walker J. Introducing wavelets and time—frequency analysis. Engineering in Medicine & Biology Magazine IEEE 28.5(2009):13–13. pmid:19775951
  27. 27. Liang M., et al. Eye fundus vessel segmentation technology based on SVM. Applied Science and Technology (2017). 44, 67–71.
  28. 28. Orlando J.I.; Prokofyeva E.; Blaschko M.B. A discriminatively trained fully connected conditional random field model for blood vessel segmentation in fundus images. IEEE Trans. Biomed. Eng. 2017, 64, 16–27. pmid:26930672
  29. 29. Zhang J, Chen Y, Bekkers E, et al. Retinal vessel delineation using a brain-inspired wavelet transform and random forest. Pattern Recognit. 2017, 69, 107–123.
  30. 30. Wu Yicheng, et al. Multiscale Network Followed Network Model for Retinal Vessel Segmentation. International Conference on Medical Image Computing & Computer-assisted Intervention Springer, Cham, 2018.
  31. 31. Ronneberger O., Fischer P., Brox T. U-net: convolutional networks for biomedical image segmentation. In: Navab N., Hornegger J., Wells W.M., Frangi A.F.(eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234–241. Springer, Cham (2015). https://doi.org/10.48550/arXiv.1505.04597
  32. 32. He K., Zhang X., Ren S., and Sun J., Deep residual learning for image recognition, in CVPR, 2016, pp. 770–778 (2016).
  33. 33. Staal J., et al. Ridge-based vessel segmentation in color images of the retina. " IEEE Transactions on Medical Imaging 23.4(2004):501–509. pmid:15084075
  34. 34. Fraz Muhammad Moazam, et al. Blood vessel segmentation methodologies in retinal images-a survey. Computer methods and programs in biomedicine 108.1 (2012): 407–433. pmid:22525589
  35. 35. Zhuang J. Laddernet: multi-path networks based on u-net for medical image segmentation. arXiv preprint arXiv:1810.07810 (2018).
  36. 36. Li S., Xu J., Chen R. (2020). U-Net Neural Network Optimization Method Based on Deconvolution Algorithm. In: Yang H., Pasupa K., Leung A.CS., Kwok J.T., Chan J.H., King I. (eds) Neural Information Processing. ICONIP 2020. Lecture Notes in Computer Science(), vol 12532. Springer, Cham. https://doi.org/10.1007/978-3-030-63830-6_50
  37. 37. He K., Zhang X., Ren S., and Sun J. Identity mappings in deep residual networks.In ECCV (2016). https://doi.org/10.48550/arXiv.1603.05027.
  38. 38. Hu Jie, et al.:Squeeze-and-Excitation Networks.In: CVPR, 2018, arXiv preprint arXiv:1709.01507 (2018).
  39. 39. Chen L C, Papandreou G, Kokkinos I, et al. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Transactions on Pattern Analysis & Machine Intelligence, 2016, 40(4):834–848. pmid:28463186
  40. 40. Wang Y., Zhu C., Yin XC. (2020). A Hybrid Self-Attention Model for Pedestrians Detection. In: Yang H., Pasupa K., Leung A.CS, Kwok J.T., Chan J.H., King I. (eds) Neural Information Processing. ICONIP 2020. Lecture Notes in Computer Science(), vol 12532. Springer, Cham. https://doi.org/10.1007/978-3-030-63830-6_6
  41. 41. Wang T. et al. A Stagewise Refinement Model for Detecting Salient Objects in Images. in Proc. ICCV, Venice, Italy, Oct. 2017, pp, 2380–7504.
  42. 42. Orujov F., et al. "Fuzzy based image edge detection algorithm for blood vessel detection in retinal images." Applied Soft Computing (2020).
  43. 43. Roychowdhury S, Koozekanani DD, Parhi KK. Iterative Vessel Segmentation of Fundus Images. IEEE Transacions on Biomedical Engineering. 2015; 62:1738–1749. pmid:25700436
  44. 44. Azzopardi G., Strisciuglio N., Vento M., and Petkov N., Trainable COSFIRE fifilters for vessel delineation with application to retinal images, Med. Image Anal., vol. 19, no. 1, pp. 46–57, 2015. pmid:25240643
  45. 45. Zhang J., et al. Robust retinal vessel segmentation via locally adaptive derivative frames in orientation scores. IEEE Transactions on Medical Imaging 35.12(2016):1–1. pmid:27514039
  46. 46. Fraz M. M. et al., An ensemble classification-based approach applied toretinal blood vessel segmentation, IEEE Trans. Biomed. Eng., vol. 59,no. 9, pp. 2538–2548, Sep. 2012. pmid:22736688
  47. 47. Li Q. et al., A Cross-Modality Learning Approach for Vessel Segmentation in Retinal Images. IEEE Transactions on Medical Imaging 35.1(2016):109–118. pmid:26208306
  48. 48. Hu K., et al. Retinal vessel segmentation of color fundus images using multiscale convolutional neural network with an improved cross-entropy loss function. Neuro computing 2018, 309, 179–191.
  49. 49. Huang G., et al. Multi-scale dense convolutional networks for efficient prediction. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–14. https://doi.org/10.48550/arXiv.1703.09844
  50. 50. Krizhevsky A., Sutskever I., and Hinton G. ImageNet classification with deep convolutional neural networks. In Proceedings of the International Conference Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 60, pp. 1097–1105.
  51. 51. Zhang D. et al. (2020). Deep Patch-Based Human Segmentation. In: Yang H., Pasupa K., Leung A.CS, Kwok J.T., Chan J.H., King I. (eds) Neural Information Processing. ICONIP 2020. Lecture Notes in Computer Science(), vol 12532. Springer, Cham. https://doi.org/10.1007/978-3-030-63830-6_20
  52. 52. Guo C., Szemenyei M., Yi Y., Zhou W., Bian H. (2020). Residual Spatial Attention Network for Retinal Vessel Segmentation. In: Yang H., Pasupa K., Leung A.CS., Kwok J.T., Chan J.H., King I (eds) Neural Information Processing. ICONIP 2020. Lecture Notes in Computer Science(), vol 12532. Springer, Cham. https://doi.org/10.1007/978-3-030-63830-6_43
  53. 53. Wang B., Qiu S., He H. Dual Encoding U-Net for Retinal Vessel Segmentation. In: Shen D. et al. (eds) Medical Image Computing and Computer Assisted Intervention-MICCAI 2019.
  54. 54. Wu Y. et al. Vessel-Net: Retinal Vessel Segmentation Under Multi-path Supervision. In: Shen D. et al. (eds) Medical Image Computing and Computer Assisted Intervention-MICCAI 2019. MICCAI 2019. Lecture Notes in Computer Science, vol 11764. Springer, Cham. (2019). https://doi.org/10.1007/978-3-030-32239-7_30
  55. 55. Hu K., et al. Retinal vessel segmentation of color fundus images using multiscale convolutional neural network with an improved cross-entropy loss function. Neuro computing 2018, 309, 179–191.
  56. 56. Huang G., et al. Multi-scale dense convolutional networks for efficient prediction. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–14. https://doi.org/10.48550/arXiv.1703.09844
  57. 57. Krizhevsky A., Sutskever I., and Hinton G. ImageNet classification with deep convolutional neural networks. In Proceedings of the International Conference Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 60, pp. 1097–1105.
  58. 58. Zhang D. et al. (2020). Deep Patch-Based Human Segmentation. In: Yang H., Pasupa K., Leung A.CS., Kwok J.T., Chan J.H., King I. (eds) Neural Information Processing. ICONIP 2020. Lecture Notes in Computer Science(), vol 12532. Springer, Cham. https://doi.org/10.1007/978-3-030-63830-6_20
  59. 59. Guo C., Szemenyei M., Yi Y., Zhou W., Bian H. (2020). Residual Spatial Attention Network for Retinal Vessel Segmentation. In: Yang H., Pasupa K., Leung A.CS., Kwok, J.T., Chan, J.H., King, I. (eds) Neural Information Processing. ICONIP 2020. Lecture Notes in Computer Science(), vol 12532. Springer, Cham. https://doi.org/10.1007/978-3-030-63830-6_43