ArticlePDF Available

A systematic review of retinal fundus image segmentation and classification methods using convolutional neural networks

Authors:
  • Lagos State University of Science and Technology
Healthcare Analytics 4 (2023) 100261
Available online 22 September 2023
2772-4425/© 2023 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
A systematic review of retinal fundus image segmentation and classication
methods using convolutional neural networks
Ademola E. Ilesanmi
a
,
*
, Taiwo Ilesanmi
b
, Gbenga A. Gbotoso
c
a
University of Pennsylvania, 3710 Hamilton Philadelphia, PA, 19104, United States
b
National Population Commission, Abuja, Nigeria
c
Lagos State University of Science and Technology, Ikorodu, Nigeria
ARTICLE INFO
Handling Editor: Madijd Tavana
Keywords:
Convolutional neural network
Retinal fundus
Image segmentation and classication
Eye-related disorders
Retinal disease
Computer-aided image detection
ABSTRACT
Retinal fundus images play a crucial role in the early detection of eye problems, aiding in timely diagnosis and
treatment to prevent vision loss or blindness. With advancements in technology, Convolutional Neural Network
(CNN) algorithms have emerged as effective tools for recognition, delineation, and classication tasks. This study
proposes a comprehensive review of CNN algorithms used for retinal fundus image segmentation and classi-
cation. Our review follows a systematic approach, exploring diverse repositories to identify studies employing
CNN to segment and classify retinal fundus images. Utilizing CNNs in the segmentation and classication of
retinal fundus images can enhance the precision of segmentation outcomes and alleviate the sole dependence on
human experts. This approach enables more accurate segmentation results, reducing the burden on human ex-
perts. A total of sixty-two studies are included in our review, analyzing aspects such as database usage and the
advantages and disadvantages of the methods employed. The review provides valuable insights, limitations,
observations, and future directions in the eld. Despite certain limitations, the ndings indicate that CNN al-
gorithms consistently achieve high accuracies. The comprehensive examination of the included studies sheds
light on the potential of CNN in retinal fundus image analysis.
1. Introduction
The prevalence of eye-related diseases has witnessed a signicant
global increase, leading to a rise in the number of individuals experi-
encing acute conditions. According to the World Health Organization
(WHO), approximately 2.2 billion people worldwide suffer from visual
impairment. Alarmingly, around 1 billion cases could have been pre-
vented if timely detection had occurred. In the United States alone, an
estimated 40 million people are affected by varying degrees of severe
eye-related diseases, with a primary focus on conditions related to the
retina, such as glaucoma, among others [1,2]. In Africa, visual impair-
ment affects approximately 26.3 million individuals, with 20.4 million
experiencing low vision and 5.9 million being blind, contributing to
15.3% of the global blind population. The leading causes of blindness
and visual impairment in this region are uncorrected refractive errors
and cataracts. While individuals over the age of 50 are more susceptible
to visual impairment and blindness, it is important to note that these
conditions can affect people of all age groups. The International Clas-
sication of Diseases 11 (2018) categorizes visual impairment into two
main groups: 1) Distance vision impairment, ranging from mild to
blindness, and 2) Near vision impairment, characterized by acuity worse
than N6 or M.08 at 40 cm.
The degree of visual impairment is inuenced by several factors,
including the availability of preventive measures and treatment options,
access to vision rehabilitation services, and the availability of basic
amenities such as inclusive buildings, transportation, and information.
The leading causes of vision impairment encompass a range of condi-
tions, including: Cataracts, Glaucoma, Diabetic retinopathy, Corneal
opacity, Trachoma, Uncorrected refractive error, Age-related macular
degeneration. The causes of visual impairment among children can vary
depending on the country and social status. In low-income countries,
cataracts often emerge as the primary cause, whereas premature reti-
nopathy is a prominent cause in middle-income countries. Uncorrected
refractive error remains a leading cause of visual impairment in children
across all countries. In the adult population, glaucoma emerges as a
signicant concern, affecting nearly 80 million people globally and
ranking among the leading causes of blindness. Glaucoma can be clas-
sied into two main types: open-angle and closed-angle glaucoma.
* Corresponding author.
E-mail address: Ademola.Ilesanmi@Pennmedicine.upenn.edu (A.E. Ilesanmi).
Contents lists available at ScienceDirect
Healthcare Analytics
journal homepage: www.elsevier.com/locate/health
https://doi.org/10.1016/j.health.2023.100261
Received 10 May 2023; Received in revised form 6 August 2023; Accepted 12 September 2023
Healthcare Analytics 4 (2023) 100261
2
Clinicians estimate that approximately 90% of affected individuals suf-
fer from open-angle glaucoma. Diagnosis of glaucoma involves pro-
cedures such as assessing neuroretinal rim loss, conducting visual eld
tests, and evaluating nerve ber characteristics [3].
The optic cup, resembling a white cup-like structure at the center of
the eye, is positioned within the optic disk. It serves as one of the
diagnostic indicators for glaucoma. Diabetic retinopathy (DR) is another
prevalent cause of visual impairment, characterized by damage to the
tissues located in the blood vessels at the back of the eye. The WHO
identies DR as a signicant contributor to blindness, ranking as the
fourth leading cause globally [4] (refer to Figs. 2 and 3). In the early
stages of DR, there are usually no noticeable symptoms or changes in
eyesight. However, if left untreated, it can result in permanent vision
impairment. A study conducted by the Center for Disease Control and
Prevention in the United States revealed that almost one-third of adults
over the age of 40 have DR. Furthermore, over one-third of
African-Americans and Mexican-Americans have diabetes, a risk factor
for developing DR [5].
Retinal fundus images are photographs that provide a direct optical
representation of the eyes internal processes. They capture various
morphological and pathological components, including blood vessels,
the macula, fovea, optic disk, hemorrhages, arterioles, venules, exu-
dates, and microaneurysms [2]. Fig. 1 illustrates a diagram of retinal
fundus images, highlighting different signs of DR, such as exudates,
microaneurysms, hemorrhages, and neovascularization.
To prevent visual loss and maintain healthy eyesight, it is advisable
to follow the recommendations outlined below: 1) Maintain stable blood
sugar levels to minimize the risk of diabetic complications that can affect
vision. 2) Be aware of your familys eye health history as certain eye
conditions can have a hereditary component. 3) Adopt a balanced and
nutritious diet that includes eye-healthy foods to support overall eye
health. 4) Maintain a healthy weight, as obesity and excessive weight
can increase the risk of various eye diseases. 5) Wear appropriate pro-
tective eyewear, such as safety goggles or sunglasses, to shield your eyes
from potential injuries and harmful UV rays. 6) Incorporate regular
exercise into your routine, as it promotes overall well-being and good
blood circulation, which is benecial for eye health. 7) Reduce or
eliminate tobacco usage, as smoking has been linked to several eye
diseases and can exacerbate existing visual impairments. 8) Practice
proper eye hygiene, including resting your eyes periodically and
ensuring clean hands and eye lenses to reduce the risk of infections and
other eye-related issues [79].
Early detection and diagnosis are crucial in safeguarding against
visual impairment and blindness. For individuals with type 1 diabetes, it
is recommended to undergo screening for diabetic retinopathy (DR)
three to ve years after the onset of diabetes, while those with type 2
diabetes should have their rst screening within one year of diagnosis
[10]. Subsequent examinations should be scheduled every six months or
one year, depending on the severity of DR. In cases where the condition
is more severe, early and more frequent examinations may be advised.
The diagnosis of visual impairment involves a comprehensive exami-
nation, including high-quality retinal photography, and a structured
follow-up process. Various techniques, such as direct and indirect
ophthalmoscopy, stereoscopic retinal fundus photography, and mydri-
atic and nonmydriatic photographs, are utilized to detect and classify
visual impairments.
The gold standard for assessing retinal fundus images is the utiliza-
tion of stereoscopic photographs in seven standard elds, covering 30
[1113]. While this approach is accurate and effective, it requires skilled
professionals to operate, and the process can be laborious and
time-consuming. In many countries, clinicians face a heavy workload
due to low doctor-to-patient ratios, leading to potential errors and delays
in diagnosis and treatment. As a result, numerous medical photographs
taken in hospitals and clinics remain unused and stored within the fa-
cilities. Fortunately, these photographs can be valuable resources for
Computer-Aided Detection systems (CADs) [14]. CAD systems assist
doctors in interpreting medical images by processing digital images and
identifying signicant areas that may indicate potential disease spots.
Convolutional neural networks (CNNs) have been developed to
automatically detect and diagnose ophthalmic issues in retinal fundus
images. A notable advantage of CNNs is their ability to learn complex
features automatically and translate them into meaningful results. By
reducing manual procedures, CNN systems can enhance healthcare
practices and serve as a valuable second interpreter. They have the
potential to automate eye screening, detect abnormalities, and provide
exibility for healthcare practitioners. The objective of this study is to
present the ndings of various CNN algorithms used for automatic
segmentation, classication, and disease detection in retinal fundus
images. This review focuses on recent trends from 2015 onwards,
analyzing the performance of different CNN methods, databases, and
validation metrics. It also explores the roadmap and challenges associ-
ated with CNN detection and diagnosis. Previous research, such as
reference [2], has examined deep learning for retinal fundus images, but
this review aims to investigate the latest trends, including the pathology
and morphologies of retinal fundus images using CNN methods. Addi-
tionally, it will address various validation metrics, limitations, and
future directions for the segmentation and classication of retinal
fundus images. The contributions of this review can be summarized as
follows.
1. Analysis of different CNN methods for the segmentation and classi-
cation of retinal fundus images, along with an examination of
various databases and validation metrics.
2. Discussion of potential challenges and roadmap for CNN utilization
in the segmentation and classication of retinal fundus images.
The remainder of the paper is organized as follows: Section 2 pro-
vides background information and reviews methods for the
Fig. 1. Retinal fundus images showing different signs of the DR [6].
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
3
classication and segmentation of fundus images using CNN. Section 3
discusses different evaluation metrics employed in the classication and
segmentation processes. Databases are examined in Section 4, while
Section 5 presents statistical analyses of different algorithms. Finally,
Section 6 concludes the paper.
2. Background
In the 1860s, doctors began exploring potential solutions for
capturing images of the eye. By the 1880s, a partial solution had been
developed, allowing doctors to take pictures by placing a camera on the
patients head. However, they still had to wait approximately 3 min for
the lm to develop. Despite its simplicity, this development marked the
rst time that images of the eye could be captured by anyone. In 1926,
the rst fundus camera was invented, enabling photography of a portion
of the eye. It took another 70 years for the development of a retinal
fundus camera capable of capturing images spanning 130. Eventually,
in the 21st century, a non-invasive camera was introduced, capable of
capturing a 200-degree view [15].
In the realm of computers, the rst algorithm executed on a machine
was created by Ada Lovelace in 1843. Since then, numerous algorithms
have been developed to perform specic tasks. The eld of computer
vision emerged in 1966 when computers were employed to identify
objects. Mathematical analysis and quantitative applications were
introduced to computer vision in the subsequent decade, incorporating
techniques such as scale-space representation [16], contour models
[17], and Markov random elds [18]. Marvin Minskys research in
mimicking the human brain paved the way for computers to process
Fig. 2. Cases of vision loss by Countries [1].
Fig. 3. Global blindness gure [1].
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
4
information for decision-making. In 1959, Russell Kirsch invented a
digital image scanner capable of transforming images into digital data.
Lawrence Robert processed 3D information about solid objects from 2D
photographs in 1963. Kunishiko Fukushima developed the precursor of
modern CNNs in 1980. In 1999, David Lowe described a visual recog-
nition system utilizing local invariant features. In 2001, the rst
real-time face detection framework was introduced.
The breakthrough moment in computer vision occurred in 2012
when AlexNet won the ImageNet competition. Since then, numerous
researchers have utilized CNN methods for segmenting and classifying
medical images, particularly retinal fundus images. In recent years, the
U-shaped Network has emerged as a prominent and effective approach
for medical and biomedical image segmentation. Notably, the method
introduced by Ronneberger et al. [19] has gained signicant recogni-
tion. Building upon the U-Shaped architecture, several modications
and variants have been proposed, aiming to further enhance its perfor-
mance. A comprehensive discussion and exploration of these different
U-Shaped variants can be found in Ref. [20].
3. Review of methods
Numerous CNN methods have been proposed by researchers for
segmenting and classifying retinal fundus images, employing diverse
network architectures to develop sophisticated AI platforms. For
instance, Rohit Thanki [21] proposed a CNN method specically
designed to detect glaucoma in fundus images. This approach utilizes a
deep neural network for feature extraction and incorporates six machine
learning methods for classication. Promising results were obtained,
with the logistic regression algorithm achieving the highest classier
accuracy of 0.99 (99%). The method was evaluated on a dataset con-
sisting of 15 normal images and 15 images affected by glaucoma. In
another study, Phridviraj et al. [22] proposed a bi-directional Long
Short-Term Memory (LSTM) approach for the detection of DR in fundus
images. This method employs a three-fold approach. Firstly, a pre-
processing technique is applied to enhance the image quality. Secondly,
a deep learning-based efcient network is utilized to extract relevant
features. Finally, a bi-directional long-term memory classier
comprising six LSTM layers is employed to classify the images. The
proposed method achieved an accuracy of 97% when tested on three
different datasets.
In the domain of glaucoma classication in fundus images, Kamesh
Sonti and Ravindra Dhul [23] proposed a CNN-based approach. Their
method consists of 26 layers, including six convolution layers, four
pooling layers, and one fully connected layer. A softmax layer is utilized
to generate a simple mask prediction for the classier. Through
cross-validation and data augmentation techniques, the method ach-
ieved a high-performance accuracy of 96%. In a related study, Raja
Sankari et al. [24] focused on detecting retinopathy in fundus images of
preterm infants using CNN. Their approach involved preprocessing and
segmentation of the images using a modied multiresolution U-Shaped
Network [25]. Features were extracted from both traditional methods
and deep learning techniques. Subsequently, these features were fed into
the random forest algorithm for classication. These research efforts
demonstrate the application of CNN in the accurate identication and
classication of specic eye conditions, such as glaucoma and retinop-
athy. The proposed methods showcase the potential of CNN-based ap-
proaches in improving diagnosis and treatment decisions based on
retinal fundus images.
In this study we emphasize the prowess of CNN methodologies in
tackling distinct ophthalmic conditions. Our ndings underscore the
substantial potential of these techniques for precise diagnosis and
effective classication of retinal fundus images. To provide readers with
an understanding of the current trends, we have organized these CNN
methods into three categories: (1) CNN methods for classifying and
segmenting optic discs, (2) CNN methods for classifying and segmenting
arteries and veins, and (3) CNN methods for classifying and segmenting
retinal blood vessels. Each of these categories comprises a substantial
number of CNN methods that have demonstrated promising results.
Fig. 4 presents a block diagram illustrating the different CNN ap-
proaches, along with a list of prominent CNN methods used within each
category.
3.1. CNN used for segmentation and classication
3.1.1. CNN used for optic disk and optic cup
The automation of optic disk and optic cup segmentation can address
challenges encountered in the manual procedure and those anticipated
in the future. However, this segmentation technique faces various
challenges, including 1) unclear boundaries, 2) signicant variability, 3)
interference from other image components, and 4) mixed pathologies.
To overcome these challenges, researchers have proposed different
convolutional neural network (CNN) methods. Wang et al. [26] intro-
duce an encoder-decoder network that consists of two components
working in tandem. The rst component is the feature detection (FDS)
module, which preserves features by employing two stacked convolu-
tional layers (3 ×3) with batch normalization (BN) and rectied linear
unit (ReLU) activations. The second component is the cross-correction
sub-network (CCS), which reduces the impact of multiple pooling op-
erations. The decoding block, situated in the second layer of the
network, enhances contrast and combines multiple encoding features. In
a related study by Fu et al. [27], the segmentation of the optic disc (OD)
is accomplished using a combination of the U-shaped network (UNet
[19]) and probability bubbles. The images are preprocessed with an
iterative robust homomorphic surface ltering method [28].
A two-stage technique aimed at alleviating the class imbalance
constraint problem was used by Meng et al. [29]. The candidates
location was determined through a guided search procedure, and a
weighted neighborhood voting approach was utilized to generate the
localized portable position. Preprocessing of optic disk (OD) and optic
cup (OC) images plays a crucial role in medical image processing. This
signicance is evident in the study conducted by Yuna et al. [30]. In
their research, they employed the contrast-limited adaptive histogram
equalization (CLAHE) technique [31] for enhancing the image quality
before transforming it into polar coordinates [32]. For segmentation
purposes, a CNN known as W-Net, comprising feature extractor and
context extractor modules, was utilized (refer to Fig. 5 for a clearer
visualization).
Due to the intricate connections between structures within the
retinal fundus images, the segmentation of candidate regions becomes
notably complex. Wang et al. [33] and Tan et al. [34] employed both
UNet and a basic CNN for retinal fundus image segmentation. Wang
et al. utilized the UNet approach to segment candidate regions within
the vessel density map, whereas Tan et al. employed the simple CNN to
simultaneously segment and classify three distinct components: vascu-
lature, optic disk (OD), and fovea. The images were normalized and then
inputted into the CNN, which consisted of six layers including two
convolutional, two max pooling, and two fully connected layers. In order
to underscore the signicance of preprocessing in retinal fundus images,
particularly focusing on the optic disk (OD) and optic cup (OC), Veena
et al. [35] employed Gaussian ltering and image normalization tech-
niques to eliminate undesired signals. Global distribution features and
the edge histogram texture descriptor (CLAHE and Sobel edge detector
[36]) were utilized for structure analysis and detection. The Watershed
algorithm [37] was employed to extract contour shapes and localiza-
tions. Finally, the end-to-end decoder-encoder method with 39 layers
was used. In a different study, Imtiaz et al. [38] introduce a semantic
approach that integrates automated augmentation into an
encoder-decoder architecture. The encoder block consists of 18 layers
comprising 13 convolutions and 5 pooling layers, while the decoder
block consists of 20 layers, including 14 convolutions, 5 pooling layers,
and 1 softmax layer. For further details, refer to Fig. 6.
In an independent study centered on the utilization of the UNet
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
5
model, Xie et al. [39] formulated a coarse-to-ne segmentation
approach aimed at establishing an initial segmentation boundary. The
CNN methods, specically a fully convolutional network, were com-
bined with the Viterbi algorithm to segment boundaries. The network
consists of three blocks: the encoder, decoder, and sequence decoder.
The rst two blocks involve traditional encoder and decoder layers,
while the sequence encoder layers comprise a gateway module and
cascaded gate units. The gateway module receives inputs from each
decoder layer and consists of three upsampling and three convolution
layers followed by sigmoid activation, while the gate unit comprises
ReLU and softmax layers. The Viterbi algorithm [40,41,42] was
employed to decode the output of the sequence decoder, modeling the
interaction of prediction and spatial constraints. Sadhukhan et al. [43]
introduced the attention-based fully connected CNN (AFCNN). This
network consists of 19 layers, including three attention blocks, 12
convolution layers, two dropout layers, and one softmax layer. In
Priyanka et al. [44], the zero-phase component analysis was applied to
the OD image, which was then augmented and fed into the CNN. After
being segmented with the CNN, the images were further segmented
using the fuzzy c-means method [45]. Similarly, Raja et al. [46] intro-
duced a nine-layer CNN architecture comprising three convolutional
layers, four ReLU layers, two max-pooling layers, one fully connected
layer, and one softmax layer.
Chowdhury et al. [47] developed a multiscale encoder/deco-
dernguided attention network for multicomponent segmentation of
retinal and optic discs. This network incorporates a self-attention
mechanism and extracts multiscale features. Hervella et al. [48] intro-
duced a pixel-level and image-level multi-task approach for simulta-
neously classifying glaucoma and segmenting the optic disc. Gupta et al.
[49] employ the U-Net architecture for segmentation and utilize the
mayy optimization kernel extreme learning [50] for classication.
Veena et al. [51] developed an enhanced deep learning method for
segmenting the optic disc and optic cup. Xiong et al. [52] introduced the
Bayesian U-Net Hough transform annotation method, which integrates a
Bayesian variant and the copy and crop framework to enhance the
weighting of the U-Net. In a different work, Maiti et al. [53] constructed
an encoder/decoder network comprising seven subnetworks and inte-
grated a long short-term memory framework. Moreover, multiple other
investigations, including Priyanka et al. [44] and Raja et al. [46],
employed CNNs for optic disc (OD) segmentation within retinal fundus
images. In a recent 2023 study by Rajarshi Bhattacharya et al. [54], an
encoder-decoder network was augmented with an auxiliary pyramid
decoder. This network features receptive blocks, a modied attention
block, and a spatial pyramid network. Notably, promising outcomes
were obtained, with dice scores of 94% and 95% achieved for the optic
disc and optic cup segmentation, respectively. The study referenced in
Fig. 4. Categories of CNN methods for retinal fundus images.
Fig. 5. Multi-scale CNN [30].
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
6
[55] employed a single-shot multibox detector comprising three key
components: a resize module, a feature extractor, and additional feature
blocks. A comparison of different methods in this section can be found in
Tables 1 and 2.
3.1.2. CNN used for arteries and vein
In the study by Girarda et al. [59], they put forth an encoder-decoder
CNN model aimed at segmenting arteries and veins. Median ltering was
utilized for preprocessing, followed by the stacking of resultant images
into the encoder-decoder network. The network comprises 32 feature
maps and a nal convolutional layer that reduces the map to three
classes: background, arteries, and veins. Morano et al. [60] introduced a
simultaneous segmentation module inspired by the UNet method. In this
study, fundus images were preprocessed with local intensity normali-
zation and channel-wise global contrast enhancement [60]. Subse-
quently, the UNet was utilized to predict the mask. It is noteworthy that
most of the methods employed for arteries and vein segmentation in
retinal fundus images are traditional methods, as observed in the
available literature. There is a scarcity of CNN-based approaches in this
area. For a detailed comparison of different methods in this subsection,
Fig. 6. Label-based semantic segmentation [38].
Table 1
CNN used for Optic Disk and Optic Cup.
Author Ref Year CNN Name Inspiration for research procedure Accuracy
(Accuracy)
Wan et al. [26] 2021 Asymmetric deep learning network UNet [19]
M-Net [56]
Segmentation 0.937
Fu et al. [27] 2021 Fusing UNet with probability bubbles UNet Segmentation 0.99
Meng et al. [29] 2018 RGV generated CNN model LeNet-5 [57] Segmentation 0.98
Yuan et al. [30] 2021 Multi-scale W-Net M-NET and UNet Segmentation 0.95
Wang et al. [33] 2019 Coarse-to-ne deep learning UNet Segmentation 0.93
Tan et al. [34] 2017 Single convolutional neural network Multiple segmentation Segmentation 0.96
Veena et al. [35] 2021 Deep learning enhanced CNN Encoder-decoder CNN Segmentation 0.98
Imtiaz et al. [38] 2021 Label based encoder and decoder semantic
segmentation
Encoder-decoder CNN Segmentation 0.86
Xie et al. [39] 2020 SU-Net and Viterbi algorithm UNet, dilated CNN [58], Viterbi
algorithm [41,42]
Segmentation
Sadhukhan et al. [43] 2020 AFCNN FCNN Segmentation
Priyanka et al. [44] 2017 Patches CNN CNN, Fuzzy C Means Segmentation 0.95
Raja et al. [46] 2020 Traditional CNN CNN Segmentation 0.90
Chwodhung et al. [47] 2022 MGAN Attention Network Segmentation 0.91
Hervella et al. [48] 2022 Multi task method CNN Segmentation 0.95
Gupta et al. [49] 2022 MOKEL UNet Segmentation and
classication
0.89
Xiog et al. [52] 2022 Bayesian UNet UNet Segmentation 0.94
Maiti et al. [53] 2022 Multiple convolutions VGG Segmentation 0.95
Rajarshi Bhattacharya
et al.
[54] 2023 UNet Auxilliary pyramid Network UNet Segmentation 0.95
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
7
please refer to Tables 3 and 4.
3.1.3. CNN used for retinal vessel
Retinal vessel segmentation is a long-standing problem in medical
image analysis [61,62], which is accompanied by several challenges,
including.
Presence of various abnormalities of different sizes and shapes: The
presence of abnormalities surrounding the vessels in retinal fundus
(RF) images can hinder the effective segmentation of vessels.
Limited availability of annotated data: The scarcity of annotated data
can lead to overtting issues, making it a major challenge when
segmenting vessels in FR images.
Vessel structural differences: Retinal vessels exhibit variations in
thickness and structure, making it difcult to nd a single model or
network suitable for segmenting all types of vessels.
Unstructured prediction: Pixel classication differs from vessel seg-
mentation, posing challenges in predicting the vessel structure
accurately.
To tackle these challenges, several authors have employed CNN
methods for vessel segmentation in RF images. Budak et al. [63] intro-
duced the Densely Connected and Concatenated Multi-Encoder-Decoder
CNN (DCCMED-CNN). DCCMED utilizes a patch-based learning network
and includes both training and testing phases. During training, color
patches extracted from raw retina images were used as inputs, and the
network weights were trained using stochastic gradient descent methods
[64]. The network architecture consists of 2 max-pooling layers, 2 max
unpooling layers, 8 concatenated convolution layers, batch normaliza-
tion, ReLU layers, and a softmax layer. Tang et al. [65] developed the
Multi-Proportion Channel Ensemble Model (MPC-EM) for retinal vessel
segmentation. MPC-EM comprises 5 submodel networks, with each
submodel following an encoder-center-decoder structure. A center ar-
chitecture is employed as a transitional region to adjust the shape of the
feature vector. The subnetworks were optimized using triple convolu-
tional residual blocks to enhance feature extraction and alleviate the
vanishing gradient problem [66]. In another approach, Zhao et al. [67]
employed a region-based CNN consisting of four parts: 1) backbone for
feature extraction, 2) region proposal network, 3) head module for
bounding-box regression, and 4) classication for mask generation. The
ResNet [68,69] was utilized for backbone feature extraction, and a
pyramid structure [70] was adopted to consider multiple scales. The
proposed network combined multi-task CNN with 27 layers, 15 con-
volutional layers, 4 duplicated feature maps, and 2 fully connected
layers.
Czepita and Nska [71] employed a shallow UNet architecture for the
segmentation of retinal fundus images. The method comprised six
stages: 1) Phase image registration: This stage involved aligning the
spatial domain of the fundus image to the moving spatial image domain
[72]. 2) Vessel probability map generator: The vessel probability stage
utilized an encoder-decoder CNN, specically the shallow UNet model,
to extract vessels from the fundus images. 3) Postprocessing: After the
initial segmentation, postprocessing techniques were applied to rene
the results and improve the accuracy of vessel delineation. 4) Second
segmentation: A second round of segmentation was performed to further
enhance the quality of vessel segmentation. 5) Region of interest se-
lection: Relevant regions of interest were selected to focus the analysis
on specic areas of the retinal image. 6) Vessel diameter measurement:
The retinal diameters were measured using a method similar to the one
proposed by Ref. [73], specically the calculation of central venular
equivalent (CRVE). Furthermore, Sun et al. [74] presented the multipath
cascaded UNet (MCU-Net) for retinal vessel segmentation. The MCU-Net
incorporated three types of data as input: raw FFA, small-scale FFA, and
large-scale FFA. The network fused vessel features from these inputs to
generate a vascular probability map as the output. It comprised an
attention gate [75] in conjunction with a residual recurrent unit [76].
The MCU-Net encompassed two specic blocks: the renement block
and the FFA image fusion block. For a more detailed understanding,
please refer to Fig. 7.
As evident from various studies, numerous researchers have
employed the UNet architecture RF segmentation. The UNet was
employed either in its original conguration or with adaptations to the
original structure. An illustration of such a modication can be found in
the work of Zhao et al. [69], who introduced the Nested U-shaped
attention network (NUA-Net) for segmenting and classifying retinal
images. The image was rst enhanced thereafter, the green channel
images were used as the network inputs (the same as the methods in
Ref. [77]). The NUA-Net has an encoder stage and each encoder stage
has a 2 ×2 max-pooling followed by convolution with batch normali-
zation, ReLU, and dropout. Guo et al. [78] proposed the multiscale
deeply supervised network with short connection (BTS-DSN). This
network used short connections to transfer semantic information be-
tween side-output layers. Two approaches were considered: the
bottom-top short connections and the top-bottom short connections. The
key element of this network is the top-bottom, and bottom-top short
connection approaches. A switch of connectivity within layers gives the
BTS-DSN a exible procedure. Guo et al. [79] proposed the multiple
deep convolutional neural network (MDCNN) for a formulated classi-
cation and segmentation task. The MDCNN was constructed by
cascading multiple networks with the same structure.
Noh et al. [60] proposed the scale-space approximation for
multi-scale representation in CNN (SSANet). The SSANet consists of 3
Table 2
Pros and cons of CNN used for Optic Disk and Optic Cup.
Advantages Disadvantages
Directly training a model using an end-to-
end process on both source and target
framework.
Effortlessly transforms data between
different forms.
Might exhibit excessive loss.
In case of inadequate conguration, it
can generate incorrect decoding result.
It is simple to overlook crucial features
originating from the encoder.
Noise detection is readily achievable due
to the superior pixel capture
methodology.
Images containing fewer symmetrical
components are acquired more quickly.
Prone to generating deceptive results.
Results from symmetric components
are notably poor.
Table 3
CNN used for Arteries and Vein.
Author Ref Year CNN Name Inspiration for research procedure Accuracy
Girard et al. [59] 2019 Joint segmentation model UNET Segmentation 0.96
Morano et al. [60] 2021 Simultaneous segmentation UNET Segmentation 0.96
Table 4
Pros and cons of CNN used for Arteries and Vein.
Advantages Disadvantages
Demonstrates efcacy in
preserving edges.
Achieves accurate outcomes
with a reduced image count.
A low noise ratio can disrupt image edges and
generate misleading edge noise, potentially
impacting accuracy.
Errors might arise during the normalization
process.
Due to the networks shallowness, exceptional
accuracy might not be attainable.
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
8
blocks, (feature generation, feature aggregation, and inference) with 33
layers. The feature generation block has 21 layers consisting of 1
convolution layer, 3 upsampling layers, and 17 ResBlock layers. The
aggregation stage perform two key procedure: it moves input before
each upsampling in the generation block. Secondly, it accepts inputs
from the nal block of the generation using 9 layers (5 Convolutional
layers, and 4 upsampling layers). The inference block collects inputs
from the aggregation and transforms these inputs to a mask. Zhuo et al.
[80] combined the size-invariant feature maps [81] with the dense
connectivity [82] (SID
2
Net) for the segmentation and classication of
RF. The SID
2
Net has two bottleneck modules, three dense blocks, two
convolutional layers and a sigmoid layer for prediction. An ablation
experiment was carried out dividing the network into the dense network
(DNet), and DNet with size-invariant feature maps (SIDNet). Hervella
et al. [83] used the multi-instance heating regression to predict RF
image segmentation. This method predicts binary maps with the pixels
corresponding to the location and labeling of the positive class of the
ground truth. The RF images were passed to the UNET framework and
the results were interpolated back into the original RF images.
The skip-chain convolutional network is a model design integrating
skip connections or shortcuts among layers, aiming to amplify infor-
mation ow and improve feature acquisition. For the specic task of
blood vessel segmentation, a vessel-specic skip-chain convolutional
network (VSSC Net) developed by Samuel et al. was employed. The
VSSC Net operates through two stages: preprocessing and segmentation.
The preprocessing stage converts RF images to grayscale then the
adaptive fractional difference approach [84] followed by the CLAHE is
applied to the grayscale image. The segmentation stage (VSSC Net) is an
end-to-end framework that takes input images of arbitrary size pro-
ducing a probability map. VSSC Net has two components: base network
architecture and novel architecture. The base network consists of
different convolutional layers split into 4 pairs. The visual geometry
group (VGG-16) [85] was used as the base network. The attention-based
before-activation residual UNet (BSEResU-Net) [86] exploits the atten-
tion mechanism and the dropblock regularization method to reduce
overtting. The images were preprocessed by transforming RGB images
to grayscale, and the CLAHE algorithm was applied to the grayscale
image. The ResU-Net has 33 convolutional layers with 16 residual op-
erations, 2 transpose convolutional layers, 2 downsampling layers, and 1
output map. Huang et al. [87] proposed the multipath scale network
(MPS-Net) for retinal vessel segmentation. The MPS-Net is an end-to-end
network that uses one high-resolution RF input and produces a proba-
bility map with two low resolutions as output. The network has 13
multi-path scale modules, 3 convolution +ReLU, 3 Normalization +
ReLU, and 1 cropping layer. The multi-path scale module has 3 regional
paths concatenated together and arranged horizontally to produce the
output. The range entropy [88] denition was introduced to describe
vessel information of the feature maps.
Tian et al. [89] proposed the multipath CNN for RF segmentation.
This network converts the original image to low-frequency and
high-frequency images with the low-pass Gaussian lter and the
high-pass Gaussian lter. The CNN consists of a convolution down-
sampling and convolution upsampling and has 32 convolutional layers
with four blocks of 64, 128, 64, and 32. Ultimately, the outcomes of the
initial and secondary CNN were merged (fused) together to generate the
conclusive segmentation result. To nd out if there is a difference be-
tween the preprocessed image and the images without preprocessing,
Atli and Gedik [90] used the Sin-Net for the segmentation of vessels in
RF images. To preprocess, the CLAHE and the multi-scale top hat
transform (MTHT) [91] were used to enhance image contrast. The
Sin-Net architecture consists of 17 layers comprising 11 convolution
operations, 2 upsampling layers, 2 down-sampling layers, 1 output and
input layer each.
The usage of reinforcement learning in RF images is gaining promi-
nence. Guo et al. [92] used CNN with reinforcement learning to segment
vessels in RF images. The images are divided into smaller patches and
sent to CNN for training. The CNN has ve components: convolution,
pooling, dropout, fully connected, and loss function. Deep CNN [93] has
received tremendous recognition in medical image processing. As an
example, Wu et al. [94] introduced the Network Followed Network
(NFN+), which encompasses four distinct modules: 1) encoder and
decoder of the initial network, 2) encoder and decoder of the subsequent
network, 3) front group of intra-network skip connections, and 4) sec-
ond group of inter-network skip connections. In summary, the NFN +
architecture incorporates two connections involving the front and fol-
lowed networks, encompassing a total of 10 integrated components,
including convolutions, batch normalization, and dropout layers. Fully
convolutional networks (FCN) have gained relevance in tasks related to
nonmedical imaging. However, such tides are changing, Hemelings et al.
[95] used the FCN for segmenting retinal vessels in RF images. They
used the method adopted by Ref. [96] to pad the region of interest to
avoid excessive contrast enhancement at the border of the image. The
CLAHE, and noise removal methods were used to preprocess the RF and
passed to network architecture for segmentation. Boudegga et al. [97]
proposed RV-Net for vessel segmentation. This method preprocesses the
RF images by replacing the black area with an average color (see
Ref. [98]), then the image was converted to LAB. The CLAHE algorithm
is applied to the image and the channels are merged and converted back
to RGB. The preprocessed image was augmented by performing image
transformation, cropping, and patch extraction [99]. The images were
fed into the network for segmentation by the RV-Net. Wang et al. [100]
proposed the hybrid CNN and ensemble random forest (RFs) [101]
method. CNN was used for segmentation while the RF was the trainable
traditional method used for classication. The CNN has 5 layers con-
sisting of the convolution, subsampling, and fully connected layer. Hu
et al. [102] conducted a study introducing a multiscale CNN with a
cross-entropy loss function for retinal vessel segmentation. The original
RF image underwent augmentation before being fed into the network.
Fig. 7. MCU-net [74].
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
9
The multiscale CNN framework comprised four distinct stages. The rst
and second stages comprised four convolutional layers and one
max-pooling layer each. At the beginning of the network, there were a
total of 20 convolutional layers and three max-pooling layers. For a
more comprehensive understanding, please refer to Fig. 8.
Zhou et al. [103] proposed the symmetric equilibrium generative
adversarial network (SEGAN) for vessel segmentation. The SEGAN is an
end-to-end synthetic neural network that utilizes the adversarial prin-
ciple. Three fundamental principles were employed in this study:
SEGAN, multiscale feature rene block (MSFRB), and attention mech-
anism (AM) [104]. Overall, there were 13 traditional layers, 5 MSFRB,
and 5 a.m. layers. The research by Ref. [105] proposed a hybrid mul-
titask deep learning for segmenting vessels. The original image was
annotated before being fed into the deep learning algorithm. The
network has two modules: 1) multitask segmentation, 2) fusion network
module. For both modules, the improved UNet framework was adopted
for segmentation and fusion. The network is an encoder-decoder seg-
mentation consisting of 20 layers. The output of the segmentation was
passed to the fusion layer for the nal output. The deformable U-Net
(DUNet) proposed by Jin et al. [106] is a U-shaped architecture with an
encoder and decoder framework. Some of the convolutional layers in the
traditional UNet were replaced with deformable convolutional blocks.
The DUNet integrates the low-level features with high-level features.
The design was constructed with 4 convolution layers, 4 batch nor-
malizations, and 4 ReLU layers. Soomro et al. [107] proposed the strided
fully connected CNN for the segmentation of vessels in RF images. The
images were preprocessed with morphological tactics and the principal
component analysis (PCA) [108]. The network has 5 fully consecutive
convolutional blocks with sizes ranging from 16, 32, 64,128, and 256
features. There is no ablation experiment in this research.
Chala et al. [109] proposed the end-to-end improved CNN for vessel
segmentation. This network used the multi-encoder-decoder principle
and a new progressive reduction model that was integrated into the
network. The network has 4 interconnected components (multi-encoder
and parallel components, RGB-encoder and green channel encoder,
decoder component, and progressive reduction components). Data
augmentation was performed to generate more data for the network.
Sun et al. [110] proposed the contextual information enhanced UNet
(CIEU-Net) with dilated convolutional module for vessel segmentation.
The cascaded dilated module and the pyramid module were integrated
to form the segmentation network. Wu et al. [111] proposed the scale
and context-sensitive network for the segmentation of vessels. The
model consists of three modules: scale-aware feature aggregation (SFA),
adaptive feature fusion (AFF), and multi-level semantic supervision
(MSS). The SFA adjusts the receptive eld dynamically to extract fea-
tures. The AFF guides the fusion between features efciently. Satha-
nanthavathi and Indumathi [112] proposed the enhanced encoder
atrous UNet (EEA-UNet) for retinal vessel segmentation. The images
were preprocessed with the CLAHE and resized to 512 ×512.
Post-processing was done by morphological operations to remove iso-
lated false positives. The EEA-UNet is an asymmetric contraction and
expansion path that replaces all the convolutions as the atrous convo-
lution to increase the receptive eld. The contracting has 5 blocks
containing 2 atrous convolutions, batch normalization, pooling, and
ReLU layer. The atrous convolution reduces the image size without
losing the signicant features in it [113].
Yin et al. [114] proposed a U-shaped deep learning fusion network
(DF-Net) for vessel image segmentation. The method involves 4 stages:
multiscale fusion, U-shaped network, feature fusion, and classier
fusion. The original image was multiscaled by the image pyramid [115]
and constructed on a multiscale input integrated into the encoder path
for information fusion. The encoder of the network encompasses 2
convolution layers, coupled with max-pooling employing ReLU activa-
tion. The decoder, on the other hand, entails 2 convolutional layers,
succeeded by up-sampling with the number of feature maps halved. The
vessel fusion module is attached to the decoder and enhanced with the
corresponding output features. This network is combined with the
Frangi lter and a deep neural network was trained. The research by
Tang et al. [116] adopts the multiscale method. The authors developed
multi-scale channel importance sorting (MSCS) for vessel segmentation.
First, the CLAHE algorithm was used to enhance the image before it was
fed to the network. The MSCS is an encoder-decoder that consists of 3
encoders and 2 decoder blocks. Each encoder block consists of
multi-scale, channel importance, and a convolution layer. Between the
encoder and decoder, the spatial attention mechanism was used instead
of the traditional skip connection to read the output and characterize the
encoder generating the attention coefcients. The research by Ref. [117]
proposed the cascaded attention residual network (AReN-UNet). The
AReN-UNet features a unique aggregated residual module, which in-
corporates concatenated max and average pooling, along with a shared
multilayer perceptron (MLP) [118] concluded with a sigmoid layer. Shi
et al. [119] developed the multiscale dense network (MD-Net) that
makes good use of the multi-scale features and the encoder features. This
network is preprocessed with the CLAHE algorithm. A residual atrous
spatial pyramid pooling (Res-ASPP) was blended into the error frame-
work and the dense multi-level fusion merges the features in the encoder
and decoder. A squeeze and extraction (SE) block was applied to the
concatenated layer for effective feature channels. The Res-ASPP has 12
layers all of which are convolution layers with varying dimensions and
sizes. The multi-level fusion mechanism and SE block perform the fusion
procedure in the network.
Tchinda et al. [120,121] used the combination of edge detection and
neural network to segment vessels in RF images. The method used
feature vectors with eight characteristic pixels. The feature vectors
include 1) image gradient obtained with edge detection (Prewitt, Sobel,
Canny, and Gaussian [122,and123]]), 2) the Laplacian of Gaussian l-
ter, 3) morphological transformation (erosion, dilation, and top hat
ltering [124]). The cascaded feed-forward network was used for seg-
mentation. Gegundez-Arias et al. [125] proposed the simplied UNet for
the segmentation of RF images. A combination of the residual block and
batch normalization in the upsampling and downsampling layers pro-
duces the required segmentation results. The simplied UNet has 10
blocks consisting of 1 CONV_ReLU1 layer, 1 convolution layer, 3 Block2
layers, 2 Block I1, and 3 Block I2 layers. Skip connections were used to
link Block I1 and Block I2 together. Maji et al. [126] combined the
attention-based neural network with transfer learning. This research
Fig. 8. Multiscale CNN [102].
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
10
used the optimized learning method to classify and grade RF images. The
attention mechanism adds attention to pixels near the vessel. The
Gaussian lter was used to normalize color balance and illumination,
then the data was augmented. The softmax graded the health risk with
0 as bad and 2 as good. Sangeethaa and Maheswari [127] used
morphological process, thresholding, edge detection, and adaptive his-
togram for segmentation, while CNN was used for classication. After
preprocessing and segmentation, the image was fed into the trained CNN
for classication (either normal or diseased).
Muthusamy et al. [128] used the CNN recurrent network (CNN-RRN)
to segment retinal images. The image was rst preprocessed with the
median lter for denoising and smoothing. Then, the image was resized
with dual-tree complex wavelet transform and then the classication
was done using renewal networks (CNN with recurrent neural network
concept [129]). Shi et al. [130] developed a graph-based convolutional
feature aggregation network for segmenting retinal vessels. The model
consists of multilevel feature extraction, and a graph-based high and low
aggregation model. In this network, features are extracted at different
levels and passed to high and low aggregation models for preprocessing
and segmentation. Song Guo [131] proposed the cascaded guided
network for retinal vessel segmentation. This network consists of three
branches each with a scale of different sizes. Each of the branches
consists of connected multiscale multidirectional feature learning
modules to learn under the low-resolution feature and segments se-
mantic maps as output. Ren et al. [132], combined the U-Net network
and the BiFPN [133] for the segmentation of retinal vessel images. This
network is an encoder-decoder network and the BiFPN was used as the
skip between the encoder and decoder. Hang et al. [134], proposed the
improved U-Net with the skip layer replaced with a multilayer
connection (NOLnet). The NOL block consists of transpose convolutions,
batch normalization, and concatenation layers. Xu & Fan [135] pro-
posed the dual-channel asymmetric convolution neural network. This
network consists of two parts (mainsegment Net and Finsesegment Net).
The original image was divided into two thick feature maps, with each
feature map used as input to the two networks.
Zhong et al. [136]; developed a multiscale dilated convolutional
neural network (MMDC-Net) to capture information from receptive
elds. They perform a cascading procedure on the U-Net and used a
multilayer fusion (MLF) model to fuse features and lter noise. A width
attention CNN (WA_net) proposed by Ref. [137] for retinal vessel seg-
mentation was used to decompose multiple channels of feature maps
into categories. Deng & Ye [138] proposed a multi-scale attention
mechanism residual deformable convolution that combines the M-sha-
ped network with the pulse-coupled neural network. Similarly, Huang
et al. [139], proposed the cascaded self-attention U-shaped network. The
network used the U-Net and the residual self-attention model to segment
retinal fundus images. The network progressed from ne to coarse
segmentation before producing the nal mask. Pavani et al. [140],
proposed the multistage dual-path interactive renement network. This
network follows a single encoder and dual decoder style. Data
augmentation with model implementation was used to boost the
network. Karlsson & Hardarson [141] proposed a CNN based on several
interconnected U-Nets. The network was simultaneously connected and
sandwiched with global features and exceptional linear units [142]. Zhu
et al. [143], proposed an inception-based U-Net that consists of two
parts: inception downsampling (connected to the encoder) and incep-
tion upsampling (connected to the decoder). A copy framework was
added to the decoder to help rene the network. The network proposed
by Yin et al. [144], used the U-Net, feature fusion, and classication
fusion network. The network was a multiscale network used to fuse all
features after segmentation. The segmentation outputs from each
encoder were downscaled and combined to yield the nal outcome.
Khan et al. [145] introduced a novel approach called the
multi-resolution contextual network for the segmentation of vessels in
RF. This network leverages multiscale features to capture contextual
dependencies through a bi-directional context learning framework. Liu
et al. [146] conducted a study utilizing the dual attention Res2Unet
(DA-Res2UNet) for retinal vessel segmentation. In this method, the
convolution layer in the traditional UNet was replaced with the Res2-
Block and a DropBlock. The spatial attention and dual attention mech-
anisms were key components of this approach. Comparative analysis
demonstrated a segmentation accuracy of 97% when compared to other
state-of-the-art methods. In a separate study, Liu et al. [147] proposed
the attention augmented Wasserstein generative adversarial network
(AA-WIGAN) for retinal vessel segmentation. The network employed an
advanced UNet as the generator and a discriminator. Experimental re-
sults exhibited promising outcomes with a segmentation accuracy of
97%. For further details and a comprehensive comparison of different
methods, please refer to Tables 5 and 6.
4. Performance measures
The evaluation metrics used to assess the performance of various
CNN methods in the segmentation and classication of retina fundus
images include the following.
1) True Positive (TP): The number of correctly identied positive in-
stances or correctly segmented/classied abnormalities in the image.
2) False Positive (FP): The number of incorrectly identied positive
instances or incorrectly segmented/classied abnormalities in the
image.
3) False Negative (FN): The number of incorrectly missed positive in-
stances or incorrectly unsegmented/unclassied abnormalities in the
image.
4) True Negative (TN): The number of correctly identied negative
instances or correctly classied background/non-abnormal regions
in the image.
5) Intersection over union (IOU): measures the degree of overlap be-
tween the ground truth (GT) and the segmented prediction, quanti-
fying it within a range of 01. This overlap assessment, commonly
referred to as the Intersection over Union (IOU), is also synonymous
with the Jaccard index.
IOU =GT Prediction
GT Prediction (1)
6) Recall: this method characterizes the relevant portion of the positive
segmentation prediction in relation to the ground truth. It aligns with
the concept of sensitivity, indicating the proportion of positive cases
accurately classied by the method.
Recall =Tp
Tp +Fn (2)
7) Precision: This metric quanties the positive predictions in relation
to the ground truth.
Precision =Tp
Tp +Fp (3)
8) Accuracy (ACC): This measurement computes the sum of true posi-
tive and true negative, divided by the total number of positives and
negatives.
ACC =Tp +Tn
Tp +Tn +Fp +Fn (4)
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
11
9) Specicity: This term denotes the ratio of negative cases that were
accurately classied.
Specificity =Tn
Tn +Fp (5)
10) Area under the curve (AUC): his metric characterizes the classi-
ers capability to distinguish between classes. A higher value of
the Area Under the Curve (AUC) indicates superior performance.
AUC =
b
a
F(x).dx (6)
Table 5
CNN used for Retinal vessel.
Author Ref Year CNN Name Inspiration for research procedure
Budak et al. [63] 2020 Densely connected/concatenated multi-encoder-decoder
CNN
Feedforward CNN Segmentation
Tang et al. [65] 2019 Multi-proportion channel ensemble model Ensemble model Segmentation
Zhao et al. [67] 2020 RCNN-based junction renement network Masked RCNN-model [148] Segmentation
Yuan et al. [71] 2021 Shallow U-Net UNet Segmentation
Sun et al. [74] 2021 Multi-path cascaded UNet UNet Segmentation,
Zhao et al. [69] 2021 Nested U-shaped attention network UNet Segmentation
Guo et al. [78] 2019 Bottom-top and top-bottom short connection deep
supervised network
Deep supervised network Segmentation
Guo et al. [79] 2018 Multiple deep CNN Deep CNN [149] Segmentation
Noh et al. [150] 2019 Scale-space approximated CNN DRIU [151] Segmentation
Zhuo et al. [80] 2020 Size-invariant and dense connectivity network DenseNet Network [82] Segmentation
Hervella et al. [83] 2020 Multi-instance heat map regression DNN Segmentation
P. M Samuel & T. Veeramalai [152] 2021 Vessel Specic Skip chain CNN Fully convolutional networks Segmentation
D. Li & S. Rahardja [86] 2021 Attention-based before-activation residual U-Net Modied UNet Segmentation
Lin et al. [87] 2021 Multi-path scale network HR-Net [153] Segmentation
Tian et al. [89] 2020 Multi-path CNN UNet Segmentation
I. Atli & O. S. Gedik [90] 2021 Sine-Net CNN Fully CNN Segmentation
Guo et al. [92] 2018 CNN with reinforcement sample learning. Reinforcement learning Segmentation
Wu et al. [94] 2020 A network followed network. Deep CNN [93] Segmentation
Hemelings et al. [95] 2019 Fully convolutional network. UNet Segmentation
Boudegga et al. [97] 2021 RV-Net UNet, AlexNet [154], VGG Segmentation
Wang et al. [100] 2015 Features and ensemble learning CNN, and RFs Segmentation and
classication
Hu et al. [102] 2018 Multiscale CNN Richer convolutional features
[155]
Segmentation
Zhou et al. [103] 2021 Equilibrium GAN UNet, GAN [156] Segmentation
Yang et al. [105] 2021 Improved UNet UNet Segmentation
Jin et al. [106] 2019 DUNet UNet, Deformable convNet
[157]
Segmentation
Soomro et al. [107] 2019 Strided FCNN SegNet [158] Segmentation
Chala et al. [109] 2021 Improved deep CNN DCNN Segmentation
Sun et al. [110] 2021 CIEU-Net UNet Segmentation
Wu et al. [111] 2021 SCS-Net UNet Segmentation
Sathananthavathi &
Indumathi
[112] 2021 EEA UNet UNet Segmentation
Yin et al. [114] 2021 DF-Net UNet Segmentation
Tang et al. [116] 2020 MSCS UNet Segmentation
Rahman et al. [117] 2021 Cascaded AReN-UNet UNet Segmentation
Shi et al. [119] 2021 MD-Net SegNet, PSPNet [159], UNet Segmentation
Tchinda et al. [120] 2021 Classical edge detection and neural network Articial neural network Segmentation
Gegundez-Arias et al. [125] 2021 Simplied UNet UNet Segmentation
Maji & Sekh [126] 2020 Tradition method with CNN CNN Classication
Sangeethaa & Maheswari [127] 2018 Trained CNN CNN Segmentation and
Classication
Muthusamy & Tholkapiyan [128] 2019 CNN-RNN CNN Segmentation and
Classication
Shi et al. [130] 2022 Graph-based network Multilevel method Segmentation
Guo [131] 2022 Cascade guided network Multiscale method Multiscale methods
Ren et al. [132] 2022 Combined U-Net UNet Segmentation
Huang et al. [134] 2022 Improved U-Net UNet Segmentation
Xu & Fan [135] 2022 Dual semantic CNN CNN Segmentation
Zhong et al. [136] 2022 MMDC-Net UNet Segmentation
Deng et al. [138] 2022 Multiscale attention mechanism Attention network Segmentation
Huang et al. [139] 2022 Cascaded self-attention UNet Segmentation
Pavani et al. [140] 2022 Decoder/encoder network UNet Segmentation
Karlsson & Hardarson [141] 2022 CNN based U-Net UNet Segmentation
Zhu et al. [143] 2022 Inception U-Net UNet Segmentation
Yin et al. [144] 2022 Multiple networks UNet Segmentation and
classication
Renyuan Liu et al. [146] 2023 dual attention Res2Unet UNet Segmentation
Meilin Liu et al. [147] 2023 AA-WIGAN UNet Segmentation
Khan et al. [145] 2023 Multiscale context Network Reccurent Network Segmentation
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
12
Where a and b are the upper and lower limits on the curve, and F(x) is
the equation of the curve.
11) Dice measure (DM): This measurement evaluates the degree of
overlap between the predicted segmentations and the ground
truth. Higher values indicate accurate predictions.
DM =2/prediction GT/
/Prediction/+/GT/(7)
12) F1 scores: measures the function of precision and recall.
F1=2Precision Recall
Precision +Recall (8)
13) Mathew correction coefcient (MCC): measures the number of
true and false positives and negatives.
MCC =Tp Tn Fp Fn

(Tp +Fp)(Tp Fn)(Tn +Fp)(Tn +Fn)
(9)
These metrics provide insights into the accuracy and effectiveness of
the CNN methods in detecting and classifying abnormalities in retina
fundus images.
5. Datasets
This section provides an overview of the datasets utilized for seg-
menting and classifying retina fundus images. The datasets are pre-
sented in the order of their usage, including DRIVE [160], STARE [161],
CHASE-DB1 [162], HRF [163], MESSIDOR [164], IOSTAR [165],
ORIGA [166], REFUGE [166], DB1 [167], DB0 [168], Kaggle dataset,
DRISHTI-GS [169], NIVE [170], RIM-ONEr3 [171], DRIONS-DB [172],
RITE [173], WIDE [174], SYNTHE [175], LES-AV [176], RIGA [177],
DUKE, DCA, EIARG1, and AV-INSPIRE. Fig. 9 illustrates the list of
datasets with the highest usage, while datasets with a usage count of less
than 2 were excluded from the graph. The prominent datasets in terms of
usage are DRIVE, STARE, CHASE-DB1, and HRF. It is evident that a
valuable dataset should possess relevance, usability, and high-quality
attributes [178]. Therefore, we posit that datasets with greater usage
exemplify these aforementioned characteristics. In summary, the ma-
jority of databases discussed in this study are meticulously curated re-
positories of retinal fundus images. These databases represent
invaluable assets for medical research, fostering the creation of sophis-
ticated diagnostic tools and therapies to enhance both ocular health and
patient well-being.
6. Analyses
The authors conducted a comprehensive search across various online
repositories to identify research papers focused on the classication and
segmentation of retina fundus images using CNN. Initially, a total of 300
papers were gathered through this search process. Subsequently, these
papers underwent a screening process, resulting in a selection of 170
relevant papers. The authors then specically focused on papers related
to arteries, vessels, and optic discs/cups, as illustrated in Fig. 10. From
the narrowed scope, a total of 80 research papers were included in this
review, specically addressing the classication and segmentation of
veins, optic discs, and blood vessels using CNN. Among these papers, 60
were obtained from ScienceDirect, 15 from Springer, and 5 from other
sources. While the authors made efforts to include all relevant research
articles within the scope of their study, it is possible that some
Table 6
Pros and cons of CNN used for Retinal vessel.
Advantages Disadvantages Accuracy
range
Data management is executed
with a high degree of
prociency, potentially leading
to enhanced accuracy.
The networks substantial depth
equips it to manage intricate
representations adeptly.
It is skillfully constructed with
residual blocks to mitigate the
vanishing gradient issue.
This design facilitates reduced
gradients, bias, and improved
data comprehension.
Increased parameter count can
lead to information loss,
consequently elongating
network processing time.
Inability to execute
translational invariant
procedures.
Prone to generating deceptive
results.
Demands substantial
computational resources.
0.790.97
Fig. 9. Datasets used for classication and segmentation of retina fundus images.
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
13
interesting studies may have been inadvertently overlooked.
For a more in-depth examination of the data origins and the
dispersion of publications across various years, refer to Fig. 11. To
explore the contrasts in accuracy between optic disc (OD) and optic cup
(OC) segmentation, consult Fig. 12. Observing Fig. 12 reveals that the
spectrum of achieved accuracies spans from a minimum of 86% to a
maximum of 99%. Notably, a signicant proportion of accuracy scores
cluster within the range of 93%96%. These visual representations serve
to augment our understanding of the origins of the incorporated studies
as well as the chronological distribution of research ndings within the
eld.
This review encompassed a total of 2 papers focusing on arteries and
veins, 15 papers on optic disc and cup, and 58 papers on retina vessels.
Five additional research papers published in 2023 have utilized CNN for
disease detection, employing the entire retinal fundus image for anal-
ysis. Among these, 4 studies employed a combination of CNN methods
with traditional approaches, as indicated in Tables 6 and 2 Notably,
reference [126] solely concentrated on classication, whereas refer-
ences [127,128], and [100] presented segmentation and classication
approaches specically for retina fundus images. Fig. 12 provides an
analysis of the accuracy metric specically for the optic disc and cup.
However, for a comprehensive overview of accuracy and other perfor-
mance measures, referring to the tables is recommended, as they contain
additional information on various metrics across different studies.
Fig. 10. Approach for the review process.
Fig. 11. Number of papers published yearly.
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
14
7. Conclusions and future directions
The objective of this study is to provide a comprehensive review of
CNN methods used for the segmentation and classication of retinal
fundus images. These CNN algorithms have demonstrated impressive
performance, achieving approximately 99% accuracy in certain cases. In
the eld of medical practice, these CNN methods have provided valuable
insights that clinicians have utilized to draw conclusions about specic
pathologies. They have aided in disease diagnosis and the prediction of
ophthalmic conditions in patients. The overall performance of CNN al-
gorithms in this domain has positioned them as promising alternatives to
traditional methods. Notably, the IDX (Intact Dilated Extraction) tech-
nique has already gained approval for practical use in medium and large
healthcare facilities, physician groups, and medical management rms
[179,180,181]. As CNN algorithms continue to improve in stability and
efciency, their relevance is expected to be widely accepted, opening
doors for their application in other critical areas of healthcare delivery.
7.1. Observations
From the reviewed papers, it is evident that the majority of CNN
methods for retinal fundus image segmentation are built upon the
UNet, attention mechanisms, and adversarial networks. This pref-
erence can be attributed to the strong performance of these net-
works, particularly the UNet, in medical image tasks. Considering
that retinal fundus image segmentation requires a network that
accurately addresses the task, leveraging the UNet architecture is a
suitable choice.
The utilization of preprocessing methods has contributed to
improved accuracy. Many researchers have employed techniques
such as Contrast Limited Adaptive Histogram Equalization (CLAHE)
for image enhancement and various lters for denoising. While the
use of different denoising methods has not signicantly enhanced
results, the application of CLAHE for image enhancement has
demonstrated performance improvements.
Conversion of color retinal fundus images from RGB format to LAB
and subsequent conversion back to RGB after preprocessing has been
a common practice. This conversion facilitates a conducive envi-
ronment for preprocessing procedures.
Acquiring a large dataset is a challenge in medical image analysis due
to limited availability. To address this, most CNN methods have
employed data augmentation techniques as a preprocessing step to
expand the dataset.
Multilevel, multiscale, and multidirectional networks have been
prevalent in papers published in 2022. These networks involve the
combination of multiple encoder and decoder frameworks.
Retinal vessel segmentation has emerged as the most commonly
studied application of CNN methods for segmentation and classi-
cation, likely due to the prominent visibility of retinal vessels in
fundus images.
Recent trends involve incorporating attention mechanisms, Bilinear
Feature Pyramid Networks (BIFPN), or other networks into the skip
connections of the UNet architecture. Many CNN networks proposed
in 2022 have adopted this approach.
7.2. Limitations and future directions
Although CNN algorithms have made signicant contributions to
fundus image segmentation and classication, there are still challenges
that need to be addressed when applying CNN to retinal fundus (RF)
images.
1) Limited availability of labeled quality data: The quality of the
labeled data directly inuences the performance of CNN algorithms,
and obtaining a sufcient amount of accurately labeled data for RF
images remains a challenge. One reason is that experts involved in
labeling have busy schedules and may not have enough time to
accomplish the task. For instance, achieving reliable labeling for
breast ultrasound images typically requires the participation of three
experts who perform the labeling and make majority-voting de-
cisions to obtain the nal labeled image. This process is demanding,
time-consuming, and requires signicant effort. To address this
issue, researchers can explore methods like the ones described in
Ref. [182], which involve a combination of algorithms and human
supervision to generate labeled data. Additionally, weakly super-
vised methods can be employed to leverage the available labeled
data and expand its usage in the context of unlabeled data. These
approaches offer potential solutions to mitigate the scarcity of
labeled data in RF image analysis. unlabeled data.
Fig. 12. Accuracy comparison for optic disc and cup.
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
15
2) Performance of CNN methods in a real-life situation: Many of the
CNN methods evaluated in this review lack testing in real-life sce-
narios, leaving uncertainty about their optimal performance when
applied to practical applications. To address this concern, it is crucial
for analysts and computer vision engineers to establish close
collaboration with clinicians. By working hand-in-hand, these
interdisciplinary teams can ensure a gradual and informed deploy-
ment of CNN algorithms in real-life settings. This collaborative
approach allows for the integration of clinical expertise, validation
studies, and iterative improvements, thereby enhancing the reli-
ability and effectiveness of CNN algorithms when used in real-life
applications.
3) Unavailability of standard measuring metrics: The use of diverse
evaluation metrics for reporting quantitative results is evident in the
research, as these metrics capture different aspects of performance.
Consequently, authors have the exibility to choose any metric from
the available list to quantify their ndings. While this review has
addressed the issue to some extent by aggregating two common ac-
curacies from the studies, there is still room for further efforts to
minimize the proliferation of metrics. We propose the adoption of a
standardized procedure that selects three or four metrics as a
benchmark for the segmentation and classication of RF images. By
establishing this standardized set of metrics, uniformity can be
achieved in the evaluation process, allowing for better comparison
and interpretation of results across different studies.
4) CNN high computation space: The resource-intensive nature of
CNN algorithms is widely recognized, often necessitating a signi-
cant computational infrastructure. Inadequate computational re-
sources can pose challenges to achieving higher accuracies in
segmentation tasks. To address this issue, we call for the develop-
ment of additional cloud computing platforms, such as Collab, that
offer ample computational space. These platforms should be acces-
sible to researchers at minimal or no cost, facilitating the execution
of their applications and alleviating the burden of limited compu-
tational resources. By expanding the availability of such cloud
computing platforms, researchers can harness the necessary
computational power to enhance the performance of CNN algorithms
in segmentation tasks without incurring substantial expenses.
Author contributions
All authors contributed to the studys conception and design. Mate-
rial preparation and data collection were performed by AEI. Analysis
was done by AEI, TI., AG. The manuscript was written and polished by
AEI.
Funding
No funding applicable to this research.
Declaration of competing interest
The authors declare that they have no known competing nancial
interests or personal relationships that could have appeared to inuence
the work reported in this paper.
Data availability
No data was used for the research described in the article.
Acknowledgements
We would like to sincerely express our deep gratitude to our
esteemed mentors: Professor David A. Wolk from the Penn Memory
Center at the University of Pennsylvania, Professor Sandhitsu Das from
PICSL at the University of Pennsylvania, Professor Jayaram K. Udupa
and Professor Drew A. Torigian from the MIPG at the University of
Pennsylvania, and Professor Stanislav Makhanov at SIIT Thammassat
University. We also extend our heartfelt thanks to the management of
Alex Ekwueme Federal University Nigeria, as well as the anonymous
referees of the review, for their invaluable remarks and signicant
contributions.
References
[1] World Health Organization, Blindness and Vision Impairment, 2022. https
://www.who.int/news-room/fact-sheets/detail/blindness-and-visual-impairmen
t. June 19 2023.
[2] S. Sengupta, A. Singh, H.A. Leopold, T. Gulati, V. Lakshminarayanan, Ophthalmic
diagnosis using deep learning with fundus images - a critical review, Artif. Intell.
Med. 102 (2020 Jan), 101758, https://doi.org/10.1016/j.artmed.2019.101758.
Epub 2019 Nov 22. PMID: 31980096.
[3] M.T. Nicolela, J.R. Vianna, Pearls of Glaucoma Management, in: Optic Nerve:
Clinical Examination, Springer, Berlin, Heidelberg, 2016, pp. 1726.
[4] T. Aziz, A.E. Ilesanmi, C. Charoenlarpnopparut, Efcient and accurate
hemorrhages detection in retinal fundus images using smart window features,
Appl. Sci. 11 (2021) 6391, https://doi.org/10.3390/app11146391.
[5] A.S. Krolewski, J.H. Warram, L.I. Rand, A.R. Christlieb, E.J. Busick, C.R. Kahn,
Risk of proliferative diabetic retinopathy in juvenile-onset type I diabetes: a 40-yr
follow-up study, Diabetes Care 9 (5) (1986) 443452.
[6] T. Kauppi, V. Kalesnykiene, J.K. Kamarainen, L. Lensu, I. Sorri, H. Uusitalo,
H. K¨
alvi¨
ainen, J. Pietil¨
a, DIARETDB0: evaluation database and methodology for
diabetic retinopathy algorithms, Machine Vision and Pattern Recognition
Research Group, Lappeenranta University of Technology: Lappeenranta, Finland
73 (2006) 117.
[7] Age-Related Eye Disease Study Research Group, The relationship of dietary
carotenoid with vitamin A, E, and C intake with age-related macular
degeneration in a case-control study, Arch. Ophthalmol. 125 (9) (2007)
12251232.
[8] Age-Related Eye Disease Study Research Group, Risk factors associated with age-
related nuclear and cortical cataracts, Ophthalmol. Times 108 (8) (2001)
14001408.
[9] U.S. Department, Of Health and Human Services, Ofce of the Surgeon General,
The Health Consequences of Smoking: A Report of the Surgeon General,
Washington, D.C, 2004.
[10] M.J. Fowler, Microvascular and macrovascular complications of diabetes, Clin.
Diabetes 26 (2) (2008) 7782.
[11] ETDRSR. Group, Grading diabetic retinopathy from stereoscopic color fundus
photographsan extension of the modied Airlie House classi- cation: ETDRS
report number 10, Ophthalmol. Times 98 (5) (1991) 786806.
[12] C.J. Rudnisky, B.J. Hinz, M.T. Tennant, A.R. de Leon, M.D. Greve, High-
resolution stereoscopic digital fundus photography versus contact lens
biomicroscopy for the detection of clinically signicant macular edema,
Ophthalmol. Times 109 (2) (2002) 267274.
[13] Md Mohaimenul Islam, Hsuan-Chia Yang, Tahmina Nasrin Poly, Wen-Shan Jian,
Yu-Chuan (Jack) Li, Deep learning algorithms for detection of diabetic
retinopathy in retinal fundus photographs: a systematic review and meta-
analysis, Comput. Methods Progr. Biomed. 191 (2020), 105320.
[14] C. L Srinidhi, P. Aparna, J. Rajan, Recent advancements in retinal vessel
segmentation, J. Med. Syst. 41 (2017) 70, https://doi.org/10.1007/s10916-017-
0719-2.
[15] https://www.invision2020.com/retinal-imaging-history/. (Accessed 10 October
2021).
[16] Ademola Enitan Ilesanmi, Oluwagbenga Paul Idowu, Stanislav S. Makhanov,
Multiscale superpixel method for segmentation of breast ultrasound, Comput.
Biol. Med. 125 (2020), 103879.
[17] Yang Chen, Guirong Weng, An active contour model based on local pre-piecewise
tting image, Optik (2021), 168130.
[18] Daniel Lacker, Kavita Ramanan, Ruoyu Wu, Locally interacting diffusions as
Markov random elds on path space,Stochastic Processes and their Applications
140 (2021) 81114.
[19] O. Ronneberger, P. Fischer, T. Brox, U-net: convolutional networks for biomedical
image segmentation, in: Conference on Medical Image Computing and Computer
Assisted Intervention (MICCAI), 2015.
[20] N. Siddique, S. Paheding, C.P. Elkin, V. Devabhaktuni, U-net and its variants for
medical image segmentation: a review of theory and applications, IEEE Access 9
(2021) 8203182057, https://doi.org/10.1109/ACCESS.2021.3086020.
[21] Rohit Thanki, A deep neural network and machine learning approach for retinal
fundus image classication, Healthcare Analytics 3 (2023), 100140.
[22] M.S.B. Phridviraj, Raghuram Bhukya, Sujatha Madugula, Aakunuri Manjula,
Swathy Vodithala, Mohammed Sharfuddin Waseem, A bi-directional Long Short-
Term Memory-based Diabetic Retinopathy detection model using retinal fundus
images, Healthcare Analytics 3 (2023), 100174.
[23] Kamesh Sonti, Ravindra Dhul, A new convolution neural network model ‘‘KR-
NET’’ for retinal fundus glaucoma classication, Optik 283 (2023), 170861.
[24] V.M. Raja Sankari, U. Snekhalatha, Ashok Chandrasekaran, Prabhu Baskaran,
Automated diagnosis of Retinopathy of prematurity from retinal images of
preterm infants using hybrid deep learning techniques, Biomed. Signal Process
Control 85 (2023), 104883.
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
16
[25] L. Yanhong, S. Ji, Y. Lei, B. Guibin, Y. Hongnian, ResDO-Unet, A deep residual
network for accurate retinal vessel segmentation from fundus images, Biomed.
Signal Process Control 79 (1) (2023), https://doi.org/10.1016/j.
[26] Lei Wang, Juan Gu, Yize Chen, Yuanbo Liang, Weijie Zhang, Jiantao Pu,
Hao Chen, Automated segmentation of the opt ic disc from fundus images using
an asymmetric deep learning network, Pattern Recogn. 112 (2021), 107810.
[27] Yinghua Fu, Jie Chen, Li Jiang, Dongyan Pan, Xuezheng Yue, Yiming Zhu, Optic
disc segmentation by U-net and probability bubble in abnormal fundus images,
Pattern Recogn. 117 (2021), 107971.
[28] H. Narasimhaiyer, A. Can, B. Roysam, V. Stewart, H.L. Tanenbaum, A. Ma-
jerovics, H. Singh, Robust detection and classication of longitudinal changes in
color retinal fundus images for monitoring diabetic retinopathy, IEEE Trans.
Biomed. Eng. 53 (6) (2006) 10841098.
[29] Xianjing Meng, Xiaoming Xi, Yang Lu, Guang Zhang, Yilong Yin, Xinjian Chen,
Fast and effective optic disk localization based on convolutional neural network,
Neurocomputing 312 (2018) 285295.
[30] Xin Yuan, Lingxiao Zhou, Shuyang Yu, Miao Li, Xiang Wang, Xiujuan Zheng,
A multi-scale convolutional neural network with context for joint segmentation of
optic disc and cup, Artif. Intell. Med. 113 (2021), 102035.
[31] Ademola Enitan Ilesanmi, Utairat Chaumrattanakul, Stanislav S. Makhanov,
A method for segmentation of tumors in breast ultrasound images using the
variant enhanced deep learning, Biocybern. Biomed. Eng. 41 (2) (2021) 802818.
[32] H. Fu, J. Cheng, Y. Xu, D.W.K. Wong, J. Liu, X. Cao, Joint optic disc and cup
segmentation based on multi-label deep network and polar transformation, IEEE
Trans. Med. Imag. 37 (7) (2018) 15971605.
[33] Lei Wang, Liu Han, Yaling Lu, Hang Chen, Jian Zhang, Jiantao Pu, A coarse-to-
ne deep learning framework for optic disc segmentation in fundus images,
Biomed. Signal Process Control 51 (2019) 8289.
[34] Jen Hong Tan, U. Rajendra Acharya, Sulatha V. Bhandary, Chua Chua Kuang,
Segmentation of optic disc, fovea and retinal vasculature using a single
convolutional neural network, Journal of Computational Science 20 (2017)
7079.
[35] H.N. Veena, A. Muruganandham, T. Senthil Kumaran, A novel optic disc and optic
cup segmentation technique to diagnose glaucoma using deep learning
convolutional neural network over retinal fundus images, Journal of King Saud
University Computer and Information Sciences 34 (2022) 61876198.
[36] Neha Mathur, Shruti Mathur, Divya Mathur, A novel approach to improve Sobel
edge detector, Proc. Comput. Sci. 93 (2016) 431438.
[37] Lei Zhang, Lang Zou, Chuanyu Wu, Jiangming Jia, Jianneng Chen, Method of
famous tea sprout identication and segmentation based on improved watershed
algorithm, Comput. Electron. Agric. 184 (2021), 106108.
[38] Rakhshanda Imtiaz, M. Tariq, Khan, Syed saud naqvi, muhammad arsalan, syed
junaid nawaz, screening of glaucoma disease from retinal vessel images using
semantic segmentation, Comput. Electr. Eng. 91 (2021), 107036.
[39] Zhe Xie, Tonghui Ling, Yuanyuan Yang, Rong Shu, Brent J. Liu, Optic disc and
cup image segmentation utilizing contour-based transformation and sequence
labeling networks, J. Med. Syst. 44 (2020) 96.
[40] B. Xu, N. Wang, T. Chen, M. Li, Empirical Evaluation of Rectied Activations in
Convolutional Network, 2015, pp. 15, arXiv: 1505.00853vol. 2.
[41] A. Viterbi, Optimum detection and signal selection for partially coherent binary
communication, IEEE Trans. Inf. Theor. 11 (2) (Apr. 1965) 239246.
[42] A.J. Viterbi, A personal history of the Viterbi algorithm, IEEE Signal Process. Mag.
23 (4) (Jul. 2006) 120142.
[43] Sandip Sadhukhan, Arpita Sarkar, Debprasad Sinha, Goutam Kumar Ghorai,
Gautam Sarkar, Ashis K. Dhara, Attention based fully convolutional neural
network for simultaneous detection and segmentation of optic disc in retinal
fundus images, World academy of science, Engineering and Technology
International Journal of Medical and Health Sciences 14 (8) (2020).
[44] R. Priyanka, S.J. Grace Shoba, A. Brintha Therese, Segmentation of optic disc in
fundus images using convolutional neural networks for detection of glaucoma,
International Journal of Advanced Engineering Research and Science (IJAERS) 4
(Issue 5) (2017).
[45] F. Hoppner, F. Klawonn, A contribution to convergence theory of fuzzy c-means
and derivatives, IEEE Trans. Fuzzy Syst. 11 (5) (2003) 682694.
[46] C. Raja, N. Vinodhkumar, An efcient segmentation of optic disc using
convolution neural network for glaucoma detection in retinal images, European
Journal of Molecular & Clinical Medicine 7 (Issue 03) (2020).
[47] A.Z.M.E. Chowdhury, G. Mann, W.H. Morganb, A. Vukmirovic, A. Mehnert,
F. Sohel, Msganet-Rav, A multiscale guided attention network for artery-vein
segmentation and classication from the optic disc and retinal images, Journal of
Optometry 15 (2022) S58S69, https://doi.org/10.1016/j.optom.2022.11.001.
[48] ´
A.S. Hervella *, J. Rouco, J. Novo, M. Ortega, End-to-end multi-task learning for
simultaneous optic disc and cup segmentation and glaucoma classication in eye
fundus images, Appl. Soft Comput. 116 (2022), 108347.
[49] Indresh Kumar Gupta, Abha Choubey, Siddhartha Choubey, Mayy optimization
with deep learning enabled retinal fundus image classication model, Comput.
Electr. Eng. 102 (2022), 108176.
[50] K. Zervoudakis, S. Tsafarakis, A mayy optimization algorithm, Comput. Ind.
Eng. 145 (2020), 106559.
[51] H.N. Veena, A. Muruganandham, T. Senthil Kumaran, A novel optic disc and optic
cup segmentation technique to diagnose glaucoma using deep learning
convolutional neural network over retinal fundus images, Journal of King Saud
University Computer and Information Sciences 34 (2022) 61876619.
[52] Hao Xiong, Sidong Liu, Roneel V. Sharan, Enrico Coiera, Shlomo Berkovsky,
Weak label based Bayesian U-Net for optic disc segmentation in fundus images,
Artif. Intell. Med. 126 (2022), 102261.
[53] Souvik Maiti, Debasis Maji, Ashis Kumar Dhara, Gautam Sarkar, Automatic
detection and segmentation of optic disc using a modied convolution network,
Biomed. Signal Process Control 76 (2022), 103633.
[54] Rajarshi Bhattacharya, Rukhshanda Hussain, Agniv Chatterjee, Dwipayan Paul,
Saptarshi Chatterjee, Debangshu Dey, Py-Net, Rethinking segmentation
frameworks with dense pyramidal operations for optic disc and cup segmentation
from retinal fundus images, Biomed. Signal Process Control 85 (2023), 104895.
[55] A. Septiarini, H. Hamdani, E. Setyaningsih, E. Junirianto, F. Utaminingrum,
Automatic method for optic disc segmentation using deep learning on retinal
fundus images, Healthc Inform Res 29 (2) (2023 Apr) 145151, https://doi.org/
10.4258/hir.2023.29.2.145. Epub 2023 Apr 30. PMID: 37190738; PMCID:
PMC10209731.
[56] H. Fu, J. Cheng, Y. Xu, D. Wong, J. Liu, X. Cao, Joint optic disc and cup
segmentation based on multi-label deep network and polar transformation, IEEE
Trans. Med. Imag. 37 (7) (2018) 15971605.
[57] Y. LeCun, L. Bottou, Y. Bengio, P. Haffner, Gradient-based learning applied to
document recognition, Proc. IEEE 86 (11) (1998) 22782324.
[58] F. Yu, V. Koltun, Multi-scale context aggregation by dilated convolutions, in: 4th
International Conference on Learning Representations, ICLR 2016 - Conference
Track Proceedings, 2016.
[59] Fantin Girarda, Conrad Kavalec, Farida Cheriet, Joint segmentation and
classication of retinal arteries/veins from fundus images, Artif. Intell. Med. 94
(2019) 96109.
[60] Jos
e Morano,
Alvaro S. Hervella, Jorge Novo, Jos
e Rouco, Simultaneous
segmentation and classication of the retinal arteries and veins from color fundus
images, Artif. Intell. Med. 118 (2021), 102116.
[61] C. Wang, Z. Zhao, Y. Yu, Fine retinal vessel segmentation by combining Nest U-
net and patch-learning, Soft Comput. 25 (2021) 55195532, https://doi.org/
10.1007/s00500-020-05552-w.
[62] M. Kaur, A. Kamra, Detection of retinal abnormalities in fundus image using
transfer learning networks, Soft Comput. (2021), https://doi.org/10.1007/
s00500-021-06088-3.
[63] F. Girard, C. Kavalec, F. Cheriet, Joint segmentation and classication of retinal
arteries/veins from fundus images, Artif. Intell. Med. 94 (2019) 96109, https://
doi.org/10.1016/j.artmed.2019.02.004.
[64] Ümit Budak, Zafer C¨
omert, Musa Çıbuk, Abdulkadir S¸ engür, Dccmed-Net,
Densely connected and concatenated multi Encoder-Decoder CNNs for retinal
vessel extraction from fundus images, Med. Hypotheses 134 (2020), 109426.
[65] Tanga Peng, Qiaokang Lianga, Xintong Yana, Dan Zhang, Gianmarc Coppola,
Wei Sun, Multi-proportion channel ensemble model for retinal vessel
segmentation, Comput. Biol. Med. 111 (2019), 103352.
[66] J. Odstrcilik, et al., Retinal vessel segmentation by improved matched ltering:
evaluation on a new high-resolution fundus image database, IET Image Process. 7
(4) (2013) 373383.
[67] Zhao He, Yun Sun, Huiqi Li, Retinal vascular junction detection and classication
via deep neural networks, Comput. Methods Progr. Biomed. 183 (2020), 105096.
[68] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in:
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
2016, pp. 770778.
[69] Ruohan Zhao, Li Qin, Jianrong Wu, Jane You, A nested U-shape network with
multi-scale upsample attention for robust retinal vascular segmentation, Pattern
Recogn. 120 (2021), 107998.
[70] T.-Y. Lin, P. Dollar, R. Girshick, K. He, B. Hariharan, S. Belongie, Feature pyramid
networks for object detection, in: Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition, 2017, pp. 21172125.
[71] Maciej Czepita, Anna Fabija ´nska, Image processing pipeline for the detection of
blood ow through retinal vessels with subpixel accuracy in fundus images,
Comput. Methods Progr. Biomed. 208 (2021), 106240.
[72] S. Klein, Staring, M.K. Murphy, M.A. Viergever, J.P.W. Pluim, Elastix: a toolbox
for intensity based medical image registration, IEEE Trans. Med. Imag. 29 (1)
(2010) 196205, https://doi.org/10.1109/TMI.2009.2035616.
[73] L.D. Hubbard, R.J. Brothers, W.N. King, et al., Methods for evaluation of retinal
microvascular abnormalities associated with hypertension/sclerosis in the
atherosclerosis risk in communities study, Ophthalmol. Times 106 (12) (1999)
22692280 Dec, https://doi.org/10.1016/s0161-6420(99)90525-0.
[74] Gang Sun, Xiaoyan Liu, Xuefei Yu, Multi-path cascaded U-net for vessel
segmentation from fundus uorescein angiography sequential images, Comput.
Methods Progr. Biomed. 211 (2021), 106422.
[75] Michael Yeung, Evis Sala, Carola-Bibiane Schonlieb, Leonardo Rundo, Focus U-
Net, A novel dual attention-gated CNN for polyp segmentation during
colonoscopy, Comput. Biol. Med. 137 (2021), 104815.
[76] Lei Sang, Min Xu, Shengsheng Qian, Xindong Wu, Knowledge graph enhanced
neural collaborative ltering with residual recurrent network, Neurocomputing
454 (2021) 417429.
[77] J.V. Soares, J.J. Leandro, R.M. Cesar, H.F. Jelinek, M.J. Cree, Retinal vessel
segmentation using the 2-D Gabor wavelet and supervised classication, IEEE
Trans. Med. Imag. 25 (9) (2006) 12141222.
[78] Song Guo, Kai Wang, Hong Kang, Yujun Zhang, Yingqi Gao, Tao Li, Bts-Dsn,
Deeply supervised neural network with short connections for retinal vessel
segmentation, Int. J. Med. Inf. 126 (2019) 105113.
[79] Yanhui Guo, Ümit Budak, Abdulkadir ¸S engür, A novel retinal vessel detection
approach based on multiple deep convolution neural networks, Comput. Methods
Progr. Biomed. 167 (2018) 4348.
[80] Zhongshuo Zhuo, Jianping Huang, Ke Lu, Daru Pan, A size-invariant
convolutional network with dense connectivity applied to retinal vessel
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
17
segmentation measured by a unique index, Comput. Methods Progr. Biomed. 196
(2020), 105508.
[81] Y. Luo, H. Cheng, L. Yang, Size-invariant fully convolutional neural network for
vessel segmentation of digital retinal images, in: Signal and Information
Processing Association Annual Summit and Conference (APSIPA), 2016 Asia-
Pacic, IEEE, 2016, pp. 17.
[82] G. Huang, Z. Liu, L. Van Der Maaten, K.Q. Weinberger, Densely connected
convolutional networks, CVPR 1 (2017) 3.
[83] ´
Alvaro S. Hervella, Jorge Novo Jos´
eRouco, ManuelG. Penedo, Marcos Ortega,
Deep multi-instance heatmap regression for the detection of retinal vessel
crossings and bifurcations in eye fundus images, Comput. Methods Progr.
Biomed. 186 (2020), 105201.
[84] S. Pearl Mary, V. Thanikaiselvan, Unied adaptive framework for contrast
enhancement of blood vessels, Int. J. Electr. Comput. Eng. 10 (2020) 767777,
https://doi.org/10.11591/ijece.v10i1.pp767-777.
[85] K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale
image recognition, in: 3rd Int. Conf. Learn. Represent. ICLR 2015 - Conf, Track
Proc., 2015.
[86] Di Li, Susanto Rahardja, BseresU-Net, An attention-based before-activation
residual U-Net for retinal vessel segmentation, Comput. Methods Progr. Biomed.
205 (2021), 106070.
[87] Zefang Lin, Jianping Huang, Yingyin Chen, Xiao Zhang, Wei Zhao, Yong Li,
Ligong Lu, Meixiao Zhan, Xiaofei Jiang, Xiong Liang, A high resolution
representation network with multi-path scale for retinal vessel segmentation,
Comput. Methods Progr. Biomed. 208 (2021), 106206.
[88] C. Meng, K. Sun, S. Guan, Q. Wang, R. Zong, L. Liu, Multiscale dense
convolutional neural network for DSA cerebrovascular segmentation,
Neurocomputing 373 (2020) 123134.
[89] Chun Tian, Tao Fang, Yingle Fan, Wei Wu, Multi-path convolutional neural
network in fundus segmentation of blood vessels, biocybernetic and biomedical
engineering 40 (2020) 583595.
[90] Ibrahim Atli, Osman Serdar Gedik, Sine-Net: a fully convolutional deep learning
architecture for retinal blood vessel segmentation, Engineering Science and
Technology, Int. J. 24 (2021) 271283.
[91] X. Bai, F. Zhou, B. Xue, Image enhancement using multi-scale image features
extracted by top-hat transform, Opt Laser. Technol. 44 (2) (2012) 328336.
[92] Yanhui Guo, Ümit Budak, J. Lucas, Vespa, Elham Khorasani, Abdulkadir S
¸engür,
A retinal vessel detection approach using convolution neural network with
reinforcement sample learning strategy, Measurement 125 (2018) 586591.
[93] L.C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, A.L. Yuille, Deeplab: semantic
image segmentation with deep convolutional nets, atrous convolution, and fully
connected crfs, IEEE Trans. Pattern Anal. Mach. Intell. 40 (4) (2018) 834848,
https://doi.org/10.1109/tpami.2017.2699184.
[94] Yicheng Wu, Yong Xia, Yang Song, Yanning Zhang, Weidong Cai, NFN+: a novel
network followed network for retinal vessel segmentation, Neural Network. 126
(2020) 153162.
[95] Ruben Hemelings, Bart Elend, Ingeborg Stalmans, Karel Van Keer, Patrick De
Boever, Matthew B. Blaschko, Arteryvein segmentation in fundus images using a
fully convolutional network, Comput. Med. Imag. Graph. 76 (2019), 101636.
[96] J.V.B. Soares, J.J.G. Leandro, R.M. Cesar, H.F. Jelinek, M.J. Cree, Retinal vessel
segmentation using the 2-d gabor wavelet and supervised classication,
IEEETrans. Med. Imaging 25 (2006) 12141222.
[97] Henda Boudegga, Yaroub Elloumi, Mohamed Akil, Mohamed Hedi Bedoui,
Rostom Kachouri, Asma Ben Abdallah, Fast and efcient retinal blood vessel
segmentation method based on deep learning network, Comput. Med. Imag.
Graph. 90 (2021), 101902.
[98] Z. Jiang, H. Zhang, Y. Wang, S.-B. Ko, Retinal blood vessel segmentation using
fully convolutional network with transfer learning, Comput. Med. Imag. Graph.
68 (September) (2018) 115, https://doi.org/10.1016/j.
compmedimag.2018.04.005.
[99] P. Liskowski, K. Krawiec, Segmenting retinal blood vessels with deep neural
networks, IEEE Trans. Med. Imag. 35 (11) (2016) 23692380, https://doi.org/
10.1109/TMI.2016.2546227. November.
[100] Shuangling Wang, Yilong Yin, Guibao Cao, Benzheng Wei, Yuanjie Zheng,
Hierarchical retinal blood vessel segmentation based on feature and ensemble
learning, Neurocomputing 149 (2015) 708717.
[101] L. Breiman, Random forests, Mach. Learn. 45 (2001) 532.
[102] Kai Hu, Zhenzhen Zhang, Xiaorui Niu, Yuan Zhang, Chunhong Cao, Fen Xiao,
Xieping Gao, Retinal vessel segmentation of color fundus images using multiscale
convolutional neural network with an improved cross-entropy loss function,
Neurocomputing 309 (2018) 179191.
[103] Yukun Zhou, Zailiang Chen, Hailan Shen, Xianxian Zheng, Rongchang Zhao,
Xuanchu Duan, A rened equilibrium generative adversarial network for retinal
vessel segmentation, Neurocomputing 437 (2021) 118130.
[104] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A.N. Gomez, et al.,
Attention is all you need, in: International Conference on Learning
Representations, Long Beach, California, USA, 2017.
[105] Lei Yang, Huaixin Wang, Qingshan Zeng, Yanhong Liu, Guibin Bian, A hybrid
deep segmentation network for fundus vessels via deep-learning framework,
Neurocomputing 448 (2021) 168178.
[106] Qiangguo Jin, Zhaopeng Meng, Tuan D. Pham, Qi Chen, Leyi Wei, Ran Su, Dunet,
A deformable network for retinal vessel segmentation, Knowl. Base Syst. 178
(2019) 149162.
[107] Touque Ahmed Soomro, Ahmed J. Ab, Junbin Gao, Olaf Hellwich,
Lihong Zheng, Manoranjan Paul, Strided fully convolutional neural network for
boosting the sensitivity of retinal blood vessels segmentation, Expert Syst. Appl.
134 (2019) 3652.
[108] T.A. Soomro, A.J. A, J. Gao, O. Hellwich, M.U. Khan, M. Paul, L. Zheng,
Boosting sensitivity of a retinal vessel segmentation algorithm with convolutional
neural network, in: International Conference on Digital Image Computing:
Techniques and Applications (DICTA), 2017.
[109] Chala Mohamed, Benayad Nsiri, My Hachem El yous Alaoui,
Abdelmajid Soulaymani, Abdelrhani Mokhtari, Brahim Benaji Chandler, An
automatic retinal vessel segmentation approach based on Convolutional Neural
Networks, Expert Syst. Appl. 184 (2021), 115459.
[110] Muyi Sun, Kaiqi Li, Xingqun Qi, Hao Dang, Guanhong Zhang, Contextual
information enhanced convolutional neural networks for retinal vessel
segmentation in color fundus images, J. Vis. Commun. Image Represent. 77
(2021), 103134.
[111] Huisi Wu, Wei Wang, Jiafu Zhong, Baiying Lei, Zhenkun Wen, Jing Qin, Scs-Net,
A scale and context sensitive network for retinal vessel segmentation, Med. Image
Anal. 70 (2021), 102025.
[112] V. Sathananthavathi, G. Indumathi, Encoder enhanced atrous (EEA) unet
architecture for retinal blood vessel segmentation, Cognit. Syst. Res. 67 (2021)
8495.
[113] Y. Fisher, V. Koltun, Multi-scale context aggregation by dilated convolutions, in:
Proceeding of Computer Vision and Pattern Recognition ICLR 2016, 2016 arXiv:
1511.07122v3 [cs.CV] 30 Apr 2016.
[114] Pengshuai Yin, Hongmin Cai, Qingyao Wu, Df-Net, Deep fusion network for multi-
source vessel segmentation, Inf. Fusion 78 (2022) 199208.
[115] H. Zhao, J. Shi, X. Qi, X. Wang, J. Jia, Pyramid scene parsing network, in:
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition,
2017, pp. 28812890.
[116] Xianlun Tang, Bing Zhong, Jiangping Peng, Bohui Hao, Jie Li, Multi-scale channel
importance sorting and spatial attention mechanism for retinal vessels
segmentation, Applied Soft Computing Journal 93 (2020), 106353.
[117] Aamer Abdul Rahman, Birendra Biswal, P. Geetha Pavani, Shazia Hasan, M.V.
S. Sairam, Robust segmentation of vascular network using deeply cascaded AReN-
UNet, Biomed. Signal Process Control 69 (2021), 102953.
[118] R. Collobert, S. Bengio, Links between perceptrons, MLPs and SVMs, Proc. Intl
Conf. on Machine Learning (ICML) (2004).
[119] Zhengjin Shi, Tianyu Wang, Zheng Huang, Feng Xie, Zihong Liu, Bolun Wang,
Jing Xu, Md-Net, A multi-scale dense network for retinal vessel segmentation,
Biomed. Signal Process Control 70 (2021), 102977.
[120] Beaudelaire Saha Tchinda, Daniel Tchiotsop, Michel Noubom, Valerie Louis-Dorr,
Didier Wolf, Retinal blood vessels segmentation using classical edge detection
lters and the neural network, Inform. Med. Unlocked 23 (2021), 100521.
[121] J.V. Soares, J.J. Leandro, R.M. Cesar, H.F. Jelinek, M.J. Cree, Retinal vessel
segmentation using the 2-d gabor wavelet and supervised classication, IEEE
Trans. Med. Imag. 25 (9) (2006) 12141222.
[122] D. Ziou, S. Tabbone, Edge detection techniques- an overview, International
Journal of Pattern Recognition and Image Analysis 8 (4) (1998) 537559.
[123] J. Canny, A computational approach to edge detection, IEEE Trans. Pattern Anal.
Mach. Intell. 8 (1986) 679714.
[124] R.C. Gonzalez, R.E. Woods, S.L. Eddins, Digital Image Processing Using MATLAB,
Pearson/Prentice Hall, Upper Saddle River, New Jersey, 2004.
[125] Manuel E. Gegundez-Arias, Diego Marin-Santos, Manuel J. Isaac Perez-Borrero,
Vasallo-Vazquez, new deep learning method for blood vessel segmentation in
retinal images based on convolutional kernels and modied U-Net model,
Comput. Methods Progr. Biomed. 205 (2021), 106081.
[126] Debasis Maji, Arif Ahmed Sekh, Automatic grading of retinal blood vessel in deep
retinal image diagnosis, J. Med. Syst. 44 (2020) 180.
[127] S.N. Sangeethaa, P. Uma Maheswari, An intelligent model for blood vessel
segmentation in diagnosing DR using CNN, J. Med. Syst. 42 (2018) 175.
[128] Revathi Priya Muthusamy, S. Vinod, M. Tholkapiyan, Automatic detection of
abnormalities in retinal blood vessels using DTCWT, GLCM feature extractor and
CNN-rnn classier, Int. J. Recent Technol. Eng. 8 (Issue 1S4) (2019).
[129] Retinal vessel segmentation via A coarse-to-ne convolutional neural network, in:
IEEE International Conference on Bioinformatics and Biomedicine (BIBM), 2018.
[130] C. Shi, C. Xu, J. He, Y. Chen, Y. Cheng, Q. Yang, H. Qiu, Graph-based convolution
feature aggregation for retinal vessel segmentation, Simulat. Model. Pract. Theor.
121 (2022), 102653.
[131] Song Guo, Csgnet, Cascade semantic guided net for retinal vessel segmentation,
Biomed. Signal Process Control 78 (2022), 103930.
[132] Kan Ren, Longdan Chang, Minjie Wan, Guohua Gu, Qian Chen, An improved U-
net based retinal vessel image segmentation method, Heliyon 8 (2022), e11187.
[133] M. Tan, R. Pang, Q.V. Le, EfcientDet: Scalable and Efcient Object Detection,
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR),
2020, pp. 1077810787.
[134] J. Han, Y. Wang, H. Gong, Fundus Retinal Vessels Image Segmentation Method
Based on Improved U-Net, https://doi.org/10.1016/j.irbm.2022.03.001.
[135] Yanan Xu, Yingle Fan, Dual-channel asymmetric convolutional neural network for
an efcient retinal blood vessel segmentation in eye fundus images, Biocybern.
Biomed. Eng. 42 (2022) 695706.
[136] Zhong Xiang, Hongbin Zhang, Guangli Li, Donghong Ji, Do you need sharpened
details? Asking MMDC-Net: multi-layer multi-scale dilated convolution network
for retinal vessel segmentation, Comput. Biol. Med. 150 (2022), 106198.
[137] Dora E. Alvarado-Carrillo, Oscar S. Dalmau-Cede˜
no, Width attention based
convolutional neural network for retinal vessel segmentation, Expert Syst. Appl.
209 (2022), 118313.
A.E. Ilesanmi et al.
Healthcare Analytics 4 (2023) 100261
18
[138] Deng Xiangyu, Ye Jinhong, A retinal blood vessel segmentation based on
improved D-MNet and pulse-coupled neural network, Biomed. Signal Process
Control 73 (2022), 103467.
[139] Zheng Huang, Ming Sun, Yuxin Liu, Jiajun Wu, Csaunet, A cascade self-attention
u-shaped network for precise fundus vessel segmentation, Biomed. Signal Process
Control 75 (2022), 103613.
[140] Geetha Pavani, Birendra Biswal, Tapan Kumar Gandhi, Multistage Dpiref-Net, An
effective network for semantic segmentation of arteries and veins from retinal
surface, Neuroscience Informatics 2 (2022), 100074.
[141] Robert Arnar Karlsson, Sveinn Hakon Hardarson, Artery vein classication in
fundus images using serially connected U-Nets, Comput. Methods Progr. Biomed.
216 (2022), 106650.
[142] D.A. Clevert, T. Unterthiner, S. Hochreiter, Fast and accurate deep network
learning by exponential linear units (ELUs), 4th International Conference on
Learning Representations, ICLR 2016 - Conference Track Proceedings (2016)
114.
[143] Zifan Zhu, An Qing, Zhicheng Wang, Li Qian, Hao Fang, Zhenghua Huang, Ilu-
Net, Inception-Like U-Net for retinal vessel segmentation, Optik - International
Journal for Light and Electron Optics 260 (2022), 169012.
[144] Pengshuai Yin, Hongmin Cai, Qingyao Wu, Df-Net, Deep fusion network for multi-
source vessel segmentation, Inf. Fusion 78 (2022) 199208.
[145] Tariq M. Khan, Syed S. Naqvi, Robles-Kelly Antonio, Imran Razzak, Retinal vessel
segmentation via a Multi-resolution Contextual Network and adversarial learning,
Neural Network. 165 (2023) 310320.
[146] Renyuan Liu, Tong Wang, Xuejie Zhang, Xiaobing Zhou, DA-Res2UNet:
Explainable blood vessel segmentation from fundus images, Alex. Eng. J. 68
(2023) 539549.
[147] Meilin Liu, Zidong Wang, Li Han, Peishu Wu, Fuad E. Alsaadi, Nianyin Zeng, Aa-
Wgan, Attention augmented Wasserstein generative adversarial network with
application to fundus retinal vessel segmentation, Comput. Biol. Med. 158 (2023),
106874.
[148] K. He, G. Gkioxari, P. Doll´
ar, R. Girshick, Mask r-cnn, in: Proceedings of the IEEE
International Conference on Computer Vision, 2017, pp. 29612969.
[149] P. Liskowski, K. Krawiec, Segmenting retinal blood vessels with deep neural
networks, IEEE Trans. Med. Imag. 35 (2016) 23692380, https://doi.org/
10.1109/TMI.2016.2546227.
[150] Kyoung Jin Noh, Sang Jun Park, Soochahn Lee, Scale-space approximated
convolutional neural networks for retinal vessel segmentation, Comput. Methods
Progr. Biomed. 178 (2019) 237246.
[151] K. Maninis, J. Pont-Tuset, P.A. Arbel´
aez, L.V. Gool, Deep retinal image
understanding, in: Proc. Of Medical Image Computing and Computer Assisted
Intervention, MICCAI), 2016, pp. 140148, https://doi.org/10.1007/978-3-319-
46723-8_17.
[152] Pearl Mary Samuel, Thanikaiselvan Veeramalai, VSSC Net, Vessel specic skip
chain convolutional network for blood vessel segmentation, Comput. Methods
Progr. Biomed. 198 (2021), 105769.
[153] K. Sun, B. Xiao, D. Liu, J. Wang, Deep High-Resolution Representation Learning
for Human Pose Estimation, 2019, 09212 arXiv preprint arXiv:1902.
[154] A. Krizhevsky, I. Sutskever, G.E. Hinton, ImageNet classication with deep
convolutional neural networks, Commun. ACM 60 (6) (2017) 8490, https://doi.
org/10.1145/3065386. May.
[155] Y. Liu, M.-M. Cheng, X. Hu, K. Wang, X. Bai, Richer convolutional features for
edge detection, in: Proceedings of the IEEE Conference on Computer Vision and
Pattern Recognition, 2017, pp. 58725881.
[156] I.J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et
al., Generative adversarial nets, in: Neural Information Processing Systems
Conference, Canada, Montreal, Quebec, 2014.
[157] J. Dai, H. Qi, Y. Xiong, Y. Li, G. Zhang, H. Hu, Y. Wei, Deformable convolutional
networks, in: International Conference on Computer Vision, 2017, pp. 764773.
[158] Vijay Kendall Badrinarayanan, Cipolla Alex, Roberto, Segnet: a deep
convolutional encoder-decoder architecture for image segmentation, IEEE Trans.
Pattern Anal. Mach. Intell. 99 (2017), 11.
[159] H. Zhao, J. Shi, X. Qi, X. Wang, J. Jia, Pyramid scene parsing network, in: Proc
IEEE Comput Soc Conf Comput Vision Pattern Recognit, 2017, pp. 28812890.
Venice Italy.
[160] J. Staal, M.D. Abr`
amoff, M. Niemeijer, M.A. Viergever, B.V. Ginneken, Ridge-
based vessel segmentation in color images of the retina, IEEE Trans. Med. Imag.
23 (4) (2004) 501509.
[161] A. Hoover, M. Goldbaum, Locating the optic nerve in a retinal image using the
fuzzy convergence of the blood vessels, IEEE Trans. Med. Imag. 22 (8) (2003)
951958.
[162] Q. Li, B. Feng, L.P. Xie, P. Liang, H. Zhang, T. Wang, A cross-modality learning
approach for vessel segmentation in retinal images, IEEE Trans. Med. Imag. 35
(2016) 109118.
[163] J. Orlando, E. Prokofyeva, M. Blaschko, A discriminatively trained fully
connected conditional random eld model for blood vessel segmentation in
fundus images, IEEE Trans. Biomed. Eng. 64 (2017) 1627.
[164] L. Wang, H. Liu, Y. Lu, H. Chen, J. Zhang, J. Pu, A coarse-to-ne deep learn- ing
framework for optic disc segmentation in fundus images, Biomed. Signal Process
Control 51 (2019) 8289.
[165] S. Abbasi-Sureshjani, I. Smit-Ockeloen, E. Bekkers, B. Dashtbozorg, B. ter Haar
Romeny, Automatic detection of vascular bifurcations and crossings in retinal
images using orientation scores, in: International Symposium on Biomedical
Imaging (ISBI), IEEE, 2016, pp. 189192.
[166] J. Orlando, H. Fu, J. Breda, K. Keer, D. Bathula, A. Pinto, R. Fang, P. Heng, J. Kim,
J. Lee, J. Lee, X. Li, P. Liu, S. Lu, B. Murugesan, V. Naranjo, S. Phaye, S. Shankara-
narayana, A. Sikka, J. Son, A. Hengel, S. Wang, J. Wu, Z. Wu, G. Xu, Y. Xu, P. Yin,
F. Li, X. Zhang, Y. Xu, H. Bogunovic, REFUGE Challenge: a unied framework for
evaluating automated methods for glaucoma assessment from fundus
photographs, Med. Image Anal. 59 (2020), 101570.
[167] R.V.J.P.H. K¨
alvi¨
ainen , H. Uusitalo , Diaretdb1 diabetic retinopathy database and
evaluation protocol, Med. Image Underst. Anal. 20 07 (20 07) 61.
[168] T. Kauppi, V. Kalesnykiene, J.K. Kamarainen, L. Lensu, I. Sorri, H. Uusitalo,
H. K¨
alvi¨
ainen, J. Pietil¨
a, Diaretdb0: Evaluation Database and Methodology for
Diabetic Retinopathy Algorithms, Machine Vision and Pattern Recognition Re-
Search Group vol. 73, Lappeenranta University of Technology, Finland, 2006.
[169] J. Sivaswamy, S. Krishnadas, A. Chakravarty, G. Joshi, A.S. Tabish,
A comprehensive retinal image dataset for the assessment of glaucoma from the
optic nerve head analysis, JSM Biomed Imaging Data Papers 2 (1) (2015) 1004.
[170] R. Wang, L. Zheng, C. Xiong, C. Qiu, H. Li, X. Hou, B. Sheng, P. Li, Q. Wu, Retinal
optic disc localization using convergence tracking of blood vessels, Multimed.
Tool. Appl. 76 (22) (2016) 123.
[171] F. Fumero, S. Alay
on, J.L. Sanchez, J. Sigut, M. Gonzalez-Hernandez, Rim-one: an
open retinal image database for optic nerve evaluation, in: 24th International
Symposium on Computer-Based Medical Systems, CBMS), 2011, pp. 16, 2011.
[172] M. Abdullah, M. Fraz, S. Barman, Localization and Segmentation of Optic Disc in
Retinal Images Using Circular Hough Transform and Grow-Cut Algorithm, PeerJ,
2016.
[173] H. Kang, Y. Gao, S. Guo, X. Xu, T. Li, K. Wang, AVNet: a retinal artery/vein
classication network with category-attention weighted fusion, Comput. Methods
Progr. Biomed. 195 (2020), 105629.
[174] R. Estrada, C. Tomasi, S.C. Schmidler, S. Farsiu, Tree topology estimation, IEEE
Trans. Pattern Anal. Mach. Intell. 37 (8) (2015) 16881701.
[175] H. Zhao, H. Li, S. Maurerstroh, L. Cheng, Synthesizing retinal and neuronal
images with generative adversarial nets, Med. Image Anal. 49 (2018) 1426.
[176] J.I. Orlando, J.B. Breda, K. Van Keer, M.B. Blaschko, P.J. Blanco, C.A. Bulant,
Towards a glaucoma risk index based on simulated hemodynamics from fundus
images, in: International Conference on Medical Image Computing and Computer-
Assisted Intervention, Springer, 2018, pp. 6573.
[177] A.A. Almazroa, S. Alodhayb, E. Osman, E. Ramadan, M. Hummadi, M. Dlaim,
M. Alkatee, K. Raahemifar, V. Lakshminarayanan, Retinal Fundus Images for
Glaucoma Analysis: the RIGA Dataset, 2018.
[178] L.M. Koesten, E. Kacprzak, J.F.A. Tennison, E. Simperl, The trials and tribulations
of working with structured data: -a study on information seeking behavior, in:
Proceedings of the 2017 CHI Conference on Human Factors in Computing
Systems, ACM, New York, NY, USA, 2017, pp. 12771289, https://doi.org/
10.1145/3025453.3025838.
[179] Tao Li, Bo Wang, Chunyu Hu, Hong Kang, Hanruo Liu, Kai Wang, Huazhu Fu,
Applications of deep learning in fundus images: a review, Med. Image Anal. 69
(2021), 101971.
[180] Sourya Sengupta, Amitojdeep Singh, Henry A. Leopold, Tanmay Gulati,
Vasudevan Lakshminarayanan, Ophthalmic diagnosis using deep learning with
fundus images a critical review, Artif. Intell. Med. 102 (2020), 101758.
[181] A.M. David, Y. Lou, E. Ali, C. Warren, A. Ryan, J.C. Folk, N. Meindert, Improved
automated detection of diabetic retinopathy on a publicly available dataset
through integration of deep learning, Invest. Ophthalmol. Vis. Sci. 57 (13) (2016)
5200.
[182] Jayaram K. Udupa, Dewey Odhner, Liming Zhao, Yubing Tong, M. Monica,
S. Matsumoto, Krzysztof C. Ciesielski, Alexandre X. Falcao,
Pavithra Vaideeswaran, Victoria Ciesielski, Babak Saboury,
Syedmehrdad Mohammadianrasanani, Sanghun Sin, Raanan Arens, Drew
A. Torigian, Body-wide hierarchical fuzzy modeling, recognition, and delineation
of anatomy in medical images, Med. Image Anal. 18 (2014) 752771.
A.E. Ilesanmi et al.
... A clear visual depiction of the interior workings of the eye may be seen in retinal fundus pictures, which are photos. They record morphological and pathological features such as hemorrhages, arterioles, venules, exudates, microaneurysms, the macula, fovea, optic disk, and blood vessels [4]. Finding and identifying specific structures within the retinal fundus image, such as optic discs, lesions, retinal arteries, and other relevant locations, is often the goal of the segmentation [5]. ...
... Deep learning (DL) methodologies, which enable automatic learning of the linked characteristics for specific tasks instead of handmade procedures, have significantly improved the investigation of medical pictures [11]. Using convolution neural networks (CNNs) for the segmentation of retinal fundus images can reduce the reliance on human experts and improve the accuracy of segmentation results [4,12]. A sizable collection of retinal pictures with accompanying annotated vasculature maps is used to train the network. ...
Article
Full-text available
A unique method for improving the intelligent agents in retinal image processing is the proposed RAUGAN (Residual Attention UNet GAN) model. Reliability, accuracy, and delineation of retinal vessels are improved by RAUGAN, which incorporates residual attention mechanisms into the model through the use of Generative Adversarial Networks (GANs) and UNet architecture ideas. The model successfully segments retinal arteries from complicated backgrounds with a mean Intersection over Union (IOU) score of 0.915, which is impressive. Traditional approaches to segmenting the retinal images were not fully automated and hence required hours of labor of domain experts in annotation tasks. Yet other challenges were the accuracy in the previous studies. Incorporating residual attention modules allows the network to focus on salient regions while preserving contextual information, contributing to the model’s superior performance. A semantic web is to offer a formalism that supports it and offers a consensus for defining structure and semantics. The Semantic Web, which gives information a clear meaning and improves human–computer collaboration, is an extension of the existing Web rather than a separate one. The experimental findings show that RAUGAN has the potential to be a reliable and accurate tool for segmenting retinal vessels in medical picture analysis.
... To alleviate the redundant workload on ophthalmologists and achieve a more efficient diagnosis, many researchers have attempted to propose automated DR diagnosis methods including image-level and pixellevel diagnosis. Image-level diagnosis involves the classification of DR images [5], while pixel-level diagnosis involves the segmentation of DR lesions [6], specifically identifying four types of lesions with distinct DR pathological features: microaneurysms (MAs), hemorrhages (HEs), soft exudates (SEs), and hard exudates (EXs) [7]. Among numerous automated DR lesion segmentation methods, deep learning-based methods have been widely applied due to their high performance and easy implementation. ...
... where λ is the balance coefficient of the loss function, which is used to balance the influence of the two loss functions on the network's training optimization direction and is generally set to 0.5. In addition, considering that different lesions have different segmentation difficulties, inspired by RTNet [9], this paper assigns weights of [1,1,5,10] to HEs, EXs, SEs, and MAs to ensure the overall segmentation accuracy. Moreover, the determination of weights was guided by comprehensive experimentation. ...
Article
Full-text available
The early diagnosis of diabetic retinopathy (DR) can effectively prevent irreversible vision loss and assist ophthalmologists in providing timely and accurate treatment plans. However, the existing methods based on deep learning have a weak perception ability of different scale information in retinal fundus images, and the segmentation capability of subtle lesions is also insufficient. This paper aims to address these issues and proposes MLNet for DR lesion segmentation, which mainly consists of the Multi-Scale Attention Block (MSAB) and the Lesion Perception Block (LPB). The MSAB is designed to capture multi-scale lesion features in fundus images, while the LPB perceives subtle lesions in depth. In addition, a novel loss function with tailored lesion weight is designed to reduce the influence of imbalanced datasets on the algorithm. The performance comparison between MLNet and other state-of-the-art methods is carried out in the DDR dataset and DIARETDB1 dataset, and MLNet achieves the best results of 51.81% mAUPR, 49.85% mDice, and 37.19% mIoU in the DDR dataset, and 67.16% mAUPR and 61.82% mDice in the DIARETDB1 dataset. The generalization experiment of MLNet in the IDRiD dataset achieves 59.54% mAUPR, which is the best among other methods. The results show that MLNet has outstanding DR lesion segmentation ability.
... They are capable of automatically segmenting vessels without requiring annotated training data. Frangi vesselness filter is a well-known unsupervised learning method designed to enhance vessel-like structures in images 45 . Fraz et al. 46 developed an ensemble classification method that exploits a combination of boosting and random forests ML algorithms with a novel set of rotating invariant features. ...
... They are designed to automatically and adaptively learn spatial hierarchies of features, making them exceptionally suited for image recognition tasks. Their architecture, which allows for the simultaneous examination of several scales, improves the detection of both small and large blood vessels 45 . Furthermore, the FCNs have also been utilized for the same purpose 50 . ...
Article
Full-text available
In ophthalmic diagnostics, achieving precise segmentation of retinal blood vessels is a critical yet challenging task, primarily due to the complex nature of retinal images. The intricacies of these images often hinder the accuracy and efficiency of segmentation processes. To overcome these challenges, we introduce the cognitive DL retinal blood vessel segmentation (CoDLRBVS), a novel hybrid model that synergistically combines the deep learning capabilities of the U-Net architecture with a suite of advanced image processing techniques. This model uniquely integrates a preprocessing phase using a matched filter (MF) for feature enhancement and a post-processing phase employing morphological techniques (MT) for refining the segmentation output. Also, the model incorporates multi-scale line detection and scale space methods to enhance its segmentation capabilities. Hence, CoDLRBVS leverages the strengths of these combined approaches within the cognitive computing framework, endowing the system with human-like adaptability and reasoning. This strategic integration enables the model to emphasize blood vessels, accurately segment effectively, and proficiently detect vessels of varying sizes. CoDLRBVS achieves a notable mean accuracy of 96.7%, precision of 96.9%, sensitivity of 99.3%, and specificity of 80.4% across all of the studied datasets, including DRIVE, STARE, HRF, retinal blood vessel and Chase-DB1. CoDLRBVS has been compared with different models, and the resulting metrics surpass the compared models and establish a new benchmark in retinal vessel segmentation. The success of CoDLRBVS underscores its significant potential in advancing medical image processing, particularly in the realm of retinal blood vessel segmentation.
... The use of preprocessing procedures like image enhancement has contributed to increase machine learning model's output accuracy. Researchers have devised many techniques for it, like boosting contrast and brightness, contrast limited adaptive histogram equalization (CLAHE), other denoising filters artifact reduction, and minimizing noise as shown in Fig. 3 [12]. Diverse digital images with various visual characteristics can be created depending on the application of image processing algorithms. ...
... Diagnostics 2023, 13, 3364 2 of 20 blood vessels in the retina, helping ophthalmologists and medical professionals in early detection of various eye and systemic diseases [6,7]. The retinal vascular system, also known as the retinal vasculature, is a network of blood vessels located within the eye's retina. ...
Article
Full-text available
Retinal blood vessel segmentation is a valuable tool for clinicians to diagnose conditions such as atherosclerosis, glaucoma, and age-related macular degeneration. This paper presents a new framework for segmenting blood vessels in retinal images. The framework has two stages: a multi-layer preprocessing stage and a subsequent segmentation stage employing a U-Net with a multi-residual attention block. The multi-layer preprocessing stage has three steps. The first step is noise reduction, employing a U-shaped convolutional neural network with matrix factorization (CNN with MF) and detailed U-shaped U-Net (D_U-Net) to minimize image noise, culminating in the selection of the most suitable image based on the PSNR and SSIM values. The second step is dynamic data imputation, utilizing multiple models for the purpose of filling in missing data. The third step is data augmentation through the utilization of a latent diffusion model (LDM) to expand the training dataset size. The second stage of the framework is segmentation, where the U-Nets with a multi-residual attention block are used to segment the retinal images after they have been preprocessed and noise has been removed. The experiments show that the framework is effective at segmenting retinal blood vessels. It achieved Dice scores of 95.32, accuracy of 93.56, precision of 95.68, and recall of 95.45. It also achieved efficient results in removing noise using CNN with matrix factorization (MF) and D-U-NET according to values of PSNR and SSIM for (0.1, 0.25, 0.5, and 0.75) levels of noise. The LDM achieved an inception score of 13.6 and an FID of 46.2 in the augmentation step.
Conference Paper
Myopia is a common ocular disease that affects large populations in the world. This disease can lead to visual impairment due to the development of different types of myopic maculopathy. Therefore, prompt screening and intervention are necessary to prevent the further progression of myopic maculopathy to avoid vision loss. However, the manual inspection process is a time-consuming task that relies heavily on the experience of ophthalmologists. Towards advancing the state-of-the-art in automatic myopic maculopathy analysis using retinal fundus images, the “Myopic Maculopathy Analysis Challenge (MMAC) 2023” was launched. In this work, we present how semi-supervised learning methods can be applied to tackle two of the tasks proposed in the MMAC challenge: the segmentation of myopic maculopathy plus lesions, and the prediction of spherical equivalent. In particular, we have applied a pseudo-labeling approach for the segmentation task obtaining an average Dice similarity score of 0.6706, and a data-distillation procedure for the regression problem obtaining an R-square score of 0.7906. In both cases, we used the data from the other tasks of the challenge to increase the size of the original dataset with an automatically annotated dataset. This approach obtained the fourth position in both the segmentation and spherical equivalent prediction tasks.
Article
Full-text available
Objectives: The optic disc is part of the retinal fundus image structure, which influences the extraction of glaucoma features. This study proposes a method that automatically segments the optic disc area in retinal fundus images using deep learning based on a convolutional neural network (CNN). Methods: This study used private and public datasets containing retinal fundus images. The private dataset consisted of 350 images, while the public dataset was the Retinal Fundus Glaucoma Challenge (REFUGE). The proposed method was based on a CNN with a single-shot multibox detector (MobileNetV2) to form images of the region-of-interest (ROI) using the original image resized into 640 × 640 input data. A pre-processing sequence was then implemented, including augmentation, resizing, and normalization. Furthermore, a U-Net model was applied for optic disc segmentation with 128 × 128 input data. Results: The proposed method was appropriately applied to the datasets used, as shown by the values of the F1-score, dice score, and intersection over union of 0.9880, 0.9852, and 0.9763 for the private dataset, respectively, and 0.9854, 0.9838 and 0.9712 for the REFUGE dataset. Conclusions: The optic disc area produced by the proposed method was similar to that identified by an ophthalmologist. Therefore, this method can be considered for implementing automatic segmentation of the optic disc area.
Article
Full-text available
Background: Retinal and optic disc images are used to assess changes in the retinal vasculature. These can be changes associated with diseases such as diabetic retinopathy and glaucoma or induced using ophthalmodynamometry to measure arterial and venous pressure. Key steps toward automating the assessment of these changes are the segmentation and classification of the veins and arteries. However, such segmentation and classification are still required to be manually labelled by experts. Such automated labelling is challenging because of the complex morphology, anatomical variations, alterations due to disease and scarcity of labelled data for algorithm development. We present a deep machine learning solution called the multiscale guided attention network for retinal artery and vein segmentation and classification (MSGANet-RAV). Methods: MSGANet-RAV was developed and tested on 383 colour clinical optic disc images from LEI-CENTRAL, constructed in-house and 40 colour fundus images from the AV-DRIVE public dataset. The datasets have a mean optic disc occupancy per image of 60.6% and 2.18%, respectively. MSGANet-RAV is a U-shaped encoder-decoder network, where the encoder extracts multiscale features, and the decoder includes a sequence of self-attention modules. The self-attention modules explore, guide and incorporate vessel-specific structural and contextual feature information to segment and classify central optic disc and retinal vessel pixels. Results: MSGANet-RAV achieved a pixel classification accuracy of 93.15%, sensitivity of 92.19%, and specificity of 94.13% on LEI-CENTRAL, outperforming several reference models. It similarly performed highly on AV-DRIVE with an accuracy, sensitivity and specificity of 95.48%, 93.59% and 97.27%, respectively. Conclusion: The results show the efficacy of MSGANet-RAV for identifying central optic disc and retinal arteries and veins. The method can be used in automated systems designed to assess vascular changes in retinal and optic disc images quantitatively.
Article
Timely and affordable computer-aided diagnosis of retinal diseases is pivotal in precluding blindness. Accurate retinal vessel segmentation plays an important role in disease progression and diagnosis of such vision-threatening diseases. To this end, we propose a Multi-resolution Contextual Network (MRC-Net) that addresses these issues by extracting multi-scale features to learn contextual dependencies between semantically different features and using bi-directional recurrent learning to model former-latter and latter-former dependencies. Another key idea is training in adversarial settings for foreground segmentation improvement through optimization of the region-based scores. This novel strategy boosts the performance of the segmentation network in terms of the Dice score (and correspondingly Jaccard index) while keeping the number of trainable parameters comparatively low. We have evaluated our method on three benchmark datasets, including DRIVE, STARE, and CHASE, demonstrating its superior performance as compared with competitive approaches elsewhere in the literature.
Article
In this paper, a novel attention augmented Wasserstein generative adversarial network (AA-WGAN) is proposed for fundus retinal vessel segmentation, where a U-shaped network with attention augmented convolution and squeeze-excitation module is designed to serve as the generator. In particular, the complex vascular structures make some tiny vessels hard to segment, while the proposed AA-WGAN can effectively handle such imperfect data property, which is competent in capturing the dependency among pixels in the whole image to highlight the regions of interests via the applied attention augmented convolution. By applying the squeeze-excitation module, the generator is able to pay attention to the important channels of the feature maps, and the useless information can be suppressed as well. In addition, gradient penalty method is adopted in the WGAN backbone to alleviate the phenomenon of generating large amounts of repeated images due to excessive concentration on accuracy. The proposed model is comprehensively evaluated on three datasets DRIVE, STARE, and CHASE_DB1, and the results show that the proposed AA-WGAN is a competitive vessel segmentation model as compared with several other advanced models, which obtains the accuracy of 96.51%, 97.19% and 96.94% on each dataset, respectively. The effectiveness of the applied important components is validated by ablation study, which also endows the proposed AA-WGAN with considerable generalization ability.
Article
Diabetes is a common chronic disease and a major public health problem approaching epidemic proportions globally. People with diabetes are more likely to suffer from glaucoma than people without diabetes. Glaucoma can lead to loss of vision if not diagnosed at an early stage. This study proposes an intelligent computer-aided triage system with a deep neural network and machine learning to develop and analyze color retinal fundus images and classify glaucomatous retinal images. Deep features of retinal images from the fundus retinal image are extracted using a deep neural network, and the classification of features is performed and analyzed using different machine learning classifiers. Experimental results show that the combination of deep neural network and logistic regression-based classifier outperforms all existing glaucomatous triage systems, improving classification accuracy, sensitivity, and specificity.