This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Automated Fovea Detection Based on Unsupervised Retinal Vessel Segmentation Method

Meysam Tavakoli, Patrick Kelley, Mahdieh Nazar, Faraz Kalantari
To appear in: 2017 IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC)
DOI: 10.1109/NSSMIC.2017.8533061
M. Tavakoli and P. Kelley are with the Dept. of Physics, Indiana University-Purdue University, Indianapolis, IN, USA.M. Nazar is with the Biomedical Sciences Department, Shahid Beheshti University of Medical Sciences, Tehran, IranF. Kalantari is with Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX, USA
Abstract

The Computer Assisted Diagnosis systems could save workloads and give objective diagnostic to ophthalmologists. At first level of automated screening of systems feature extraction is the fundamental step. One of these retinal features is the fovea. The fovea is a small fossa on the fundus, which is represented by a deep-red or red-brown color in color retinal images. By observing retinal images, it appears that the main vessels diverge from the optic nerve head and follow a specific course that can be geometrically modeled as a parabola, with a common vertex inside the optic nerve head and the fovea located along the apex of this parabola curve. Therefore, based on this assumption, the main retinal blood vessels are segmented and fitted to a parabolic model. With respect to the core vascular structure, we can thus detect fovea in the fundus images. For the vessel segmentation, our algorithm addresses the image locally where homogeneity of features is more likely to occur. The algorithm is composed of 4 steps: multi-overlapping windows, local Radon transform, vessel validation, and parabolic fitting. In order to extract blood vessels, sub-vessels should be extracted in local windows. The high contrast between blood vessels and image background in the images cause the vessels to be associated with peaks in the Radon space. The largest vessels, using a high threshold of the Radon transform, determines the main course or overall configuration of the blood vessels which when fitted to a parabola, leads to the future localization of the fovea. In effect, with an accurate fit, the fovea normally lies along the slope joining the vertex and the focus. The darkest region along this line is the indicative of the fovea. To evaluate our method, we used 220 fundus images from a rural databse (MUMS-DB) and one public one (DRIVE). The results show that, among 20 images of the first public database (DRIVE) we detected fovea in 85% of them. Also for the MUMS-DB database among 200 images we detect fovea correctly in 83% on them.

Index Terms:
Retinal image, image processing, Radon transform, Fovea, top hat transformation, multi-overlapping window, contrast Enhancement, retinal blood vessel, Diabetic retinopathy

I Introduction

The Retinal images are widely used for screening of systemic diseases such as diabetic retinopathy (DR). Early detection and treatment of these diseases are critical to avoid vision loss and blindness. In the old fashion way of diagnosis, the ophthalmologists do all evaluations on retinal images to give a diagnostic result. Computer Assisted Diagnosis (CAD) systems could save workloads and may be used as an objective diagnostic tool for ophthalmologists [1, 3, 2].

Diabetes is a complicated disease that influences the blood vessel network throughout the body especially in the eyes. The result of this affection of diabetes on retinal blood vessels is DR. In the united state DR is the leading cause of blindness between adults [4] and in the western country [5]. In DR the microvasculature network changes and one method for treating that, is laser surgery therapy if this change detected early as soon as possible. It effect pathological changes in the retinal vessel network such as MAs, HEs, exudates as well as intraretinal microvascular abnormalities, venous beading and neovascularisations. According to [6], other than of DR, two circumstances that are diagnosed from the manifestation of blood vessels are Hypertension and arteriosclerosis. As well as field of bioinformatics is another application of blood vessel detection [7]. To modify retinal vasculature there are two methods: using opthalmoscopy and by analyzing fundus photographs. Detection of retinal vasculature network especially abnormal vessel was done by ophthalmologists, which is time consuming work and is associated with error and fatigue. As well as using non-professional health workers reduce our identification efficiency by 50% [8]. In other to increase the evaluation of the retinal images image processing techniques are required to extract suitable imformation about alteration of retinal vessel tree. Automated blood vessel segmentation is critical step in screening of some diseases like that: DR [9] assessment of the retinopathy of prematurity [10], or for determining of optic nerve head [11] and fovea citeWelfer.

Our CAD approach is to use vessel segmentation and parabola fitting to find the fovea. The fovea is the spot with the sharpest and clearest vision because it has no vessels blocking this dense center of cones, the photoreceptors that process bright light and color. So if disease or lesions sets in, the fovea can be a sensitive area that can quickly lead to severe damage to the patients’ overall vision [13]. And because the fovea is the smallest part of the macula region and this area doesn’t have a well-defined structure, the fovea is a rather more difficult part of the retina to detect than the optic nerve head (ONH).

Vessel segmentation is the a corner-stone in detection of other retinal land marks such as fovea [14, 15]. Moreover, it is critical step in screening of diabetic retinopathy (DR)[13, 16, 17, 18, 19, 20].
Before fovea and vessel detetction, the fundus image has to be preprocessed to ensure adequate level of success in detection. Here we applied two different methods top-hat transform.
There are several algorithms for retinal vessel segmentation [15, 10, 9, 6, 21, 22, 23, 24, 25, 26, 27, 28].

A model or template acquired from a set of ideal training retinal set was first used for detecting the fovea. One of the earlier methods to localize the fovea with templates was Sinthanayothin et al. [15]. They used a template of typical fovea intensities to correlate with the intensities of the subimages. The template of the typical fovea was defined to follow a Gaussian distribution with a fixed standard deviation for the pixel darkness with a total size of 40×40 pixels. The subimages were then compared to the template to compute a correlation coefficients and after passing a threshold, the candidate regions were selected. The candidate with the darkest center that matched the most plausible distance from the previously detected optic disk was ultimately determined to be the detected fovea. Singh et al [29] used a form of landmark of the retinal image that steers the program to localize the optic disk and fovea. They enhanced the contrast of the retinal image and used an intensity profile of a dark image template to compare and discern the fovea. Another method that obtained good results in detecting the fovea were in using morphological operations. Sekhar et al. [30] used morphological operations to detect the optic disk and then Hough Transform to detect the boundary of the optic disk. The detected optic disk diameter gives a region of interest (ROI) on where the macula and fovea should be. After some thresholding and morphological opening operations, the fovea center is ultimately obtained. Welfer similarly used morphological operations to identify the optic disk, the optic disk boundary and thus the diameter, then find a ROI where morphological operations and constraints were applied to the ROI until the fovea region was selected and detected. Some methods that inspired the development of our proposed methods stems from vessel identification, point distribution model and template features of the fovea. This approach by Fleming et al. [31] was used to detect the main vessels called the temporal arcade and then to use semi-circles as a template guide to detect the optic disk and fovea. Niemeijer et al. [14] used a combination of template-based feature extraction and feature regression analysis. Relying on a circular template trained from 500 images, the optic disk is compared to the template based on the image’s point distribution model of the intensity features and vascular features such as vessel thickness and orientation. A k-Nearest Neighbor (kNN) regressor on these features is used to predict the distance to the object of interest. Upon finding the optic disk, the fovea is then inferred by the optic disk location and detected from this search parameter. Another method, that Li and Chutatape [32] used, most closely matches with our proposed method. They used principal component analysis (PCA) which is a top-down strategy of finding common optic disk features from a training set. The PCA is comprised of three steps: obtaining a disk shape from highest 1% grayscale intensity levels, projection of the image’s disk shape on the mean disk space with corresponding statistical variances from a training set, and calculating the disk space to obtain the optic disk. The disk boundary is constructed from a modified active shape model (ASM) which uses 48 landmark points to compare to a point distribution model (PDM) that was learned from a training set and fits to the landmark points in the image. The main course of the vessel is also obtained from landmark points that are used to fit the parabolic curve that sets a searching area to find the fovea. All these methods rely on determining a feature that is common between the fovea, the optic disk and/or the overall retina. Like Li and Chutatape [32], we use the feature of the vessels assuming a parabolic shape to lead us to detect the fovea. The difference is in the way we segment and utilize the retinal vessels. In the following sections, we outline the images we used in the Materials section. The Methodology section describes the preprocessing and vessel segmentation steps that use Radon Transform on sub-images which are recombined to create a vessel map. The thickest vessels create a vessel landmark points which we fit a parabolic curve. The vertex is assumed to be the fovea center. We report our outcome in the Results section and examine and consider the limitations and future work in the Discussion section.

TABLE I: COMPARISON THE RESULTS OF SOME IMPORTANT METHODS IN FOVEA DETECTION
Study Method No. of Images Accuracy year
Sinthanayothin et al. [15] NN Algorithm (Template-Based Method) 112 80.4% 1999
Li and Chutatape [32] Modified Active Shape Model (Parabola fitting) 35 100% 2004
Fleming et al. [31] Main Vessel detection with Semiellipse fitting 1056 96.6% 2007
Tobin et al. [33] ONH Detection and Macula Localization 345 92.5% 2007
Singh et al. [29] Appearance-Based Method 502 96.61% 2008
Sekhar et al. [30] Hough transform and morphological operators 44 100% 2008
Niemeijer et al. [14] Point Distribution Model and kNN Regression 500 96.80% 2009
Welfer et al. [12] Selection of ROI and morphology 126 96.1% 2011

II Method

II-A Materials

For the fovea detection, two databsets, one rural and one publicly available databases, were used. The first rural database was named Mashhad University Medical Science Database (MUMS-DB). A total of 220 color images were captured which 200 were with DR and 20 normal from both left and right eyes of patientsl. The train set consists of 20 images and our test set consists of 200 images (180 images with DR and 20 images without DR). Images were taken with a field of view (FOV) of 50° under the same lighting conditions. According to the National Screening Committee standards all the images are captured by using a TOPCON (TRC-50EX), 50° Mydriatic retinal camera. The acquired image resolution is 2896×19442896\times 1944 TIFF format [34, 35]. The second database, DRIVE, consisting of 40 images with image resolution of 768×584768\times 584 pixels in which 33 cases did not have any sign of DR and 7 ones showed signs of early or mild DR with a 45° FOV. This database is divided into two sets: testing and training set, each of them containing 20 images [37].

II-B Preprocessing

The preprocessing step provides us an image with high possible vessel and background contrast and also unifies the histogram of the images. Here we applied top-hat transform [36] for preprocessing step
Although retinal images have three components (R, G, B), their green channel has the best contrast between vessel and background; so the green channel is selected as input image. Here, we also used mathematical morphology operators. Morphological operations work with two parts. The first one is the image to be processed and the second is called structure element. Erosion is used to reduce the objects in the image with the structure element, also known as the kernel. In contrast, dilation is used to increase the objects in the image. Secondary operations that depend on erosion and dilation are opening and closing operations. Opening, denoted as fbf\circ b, is applying an erosion followed by a dilation operation. The b represents the structure element. On the other hand, closing is first applying dilation then erosion. It is denoted as fbf\bullet b. Building from opening and closing operations, the top-hat transform is defined as the difference between the input image and its opening or closing. The top-hat transform is one of the important morphological operators [38]. Based on dilation and erosion, opening and closing denoted by fbf\circ b and fbf\bullet b are defined. The top-hat transform is defined as the difference between the input image and its opening. The top-hat transform includes white top-hat transform (WTH) and black top-hat transform (BTH) are defined by:

{WTH(x,y)=f(x,y)fb(x,y)BTH(x,y)=fb(x,y)f(x,y)\begin{cases}WTH(x,y)=f(x,y)-f\circ b(x,y)\\ BTH(x,y)=f\bullet b(x,y)-f(x,y)\\ \end{cases} (1)

In our preprocessing the basic idea is increasing the contrast between the vessels and background regions of the image. WTH or BTH extract bright and dim image regions corresponding to the used structure element. So, using the concept of WTH or BTH is one way of image enhancement through contrast enlarging based on top-hat transform. In the fundus images, the background brightness is not the same in the whole image. This background variation would lead to missed vessels or false vessel detection in the following steps. Moreover, in I, background is brighter than the details, however the vessels and other components are preferred to appear brighter than background. To deal with the problem, I is inverted as shown in I= 255 - I. Since we need a uniform background, to decrease the intensity variations in vessels background, we were firstly applied WTH(x,y) on image. It gave a high degree of differentiation between these features and background. A top-hat transformation was based on a “disk structure element” whose diameter was empirically found that the best compromise between the features and background. The disk diameter depended on the input image resolution. After top-hat transformation, we used contrast stretching to change the contrast or brightness of an image. The result was a linear mapping of a subset of pixel values to the entire range of grays, from the black to the white, producing an image with much higher contrast. Filtering a region is the process of applying a filter to a region of interest in an image, where a binary mask defines the region. In this situation we used an averaging filter on the result of image from last section (top-hat result) after that, we subtract the image from last section with the result of applying averaging filter. Before this section, in image result from top-hat transform, there are some variations in the image and some points like noise that without eliminate these points maybe supposed as some part of vessels, that increase false positive rate of our algorithm and after applying filter and subtraction this variation was removed. In other words, this section is applied for better removing background variation for better detecting of vessels [34]. The result of first step is shown in Fig. 1.

Refer to caption
Figure 1: (a) Fundus image from MUMS-DB (b) top-hat result and contrast stretching (c) result of subtraction of top-hat and filtered top-hat image.

II-C Main Procedure

Feature extraction is a crucial step for automated screening of systems. One of these important retinal features is the fovea. Therefore, in this work we are looking for the fovea localization. The fovea is a small fossa on the fundus, which is represented by a deep-red or red-brown color in color retinal images. It is the darkest part in most of the fundus images. Its geometrical relation to other retinal landmarks is employed to locate the fovea robustly.
To do this we need first to segment the blood vessels. Blood vessels can be described as bright curvilinear objects opposite of a darker background with unclear edges. The retinal blood vessels are non-uniform in intensity, length and width throughout the image. Our approach addresses the image locally and regionally where homogeneity of the vessel is more likely to happen. The algorithm is composed of 3 steps: Generation of sub-images (multi-overlapping windows), vessel segmentation, and fovea detection. In order to extract fovea, it should be extracted in local windows.

II-C1 Multi-overlapping window

In the proposed algorithm, each image was partitioned into overlapping widows in the first step. To find objects on border of sub-images, an overlapping pattern of sliding windows was defined. For determining the size of each sub-image or sliding window our knowledge database was used. In this regard, minimum and maximum sizes of targeted object specify the size of the windows (n) [36]. The n has a direct effect on the extraction accuracy.

Refer to caption
Figure 2: Window size example.

II-C2 Vessel Segmentation

In this study a Radon transform-based algorithm is proposed for detection of vessel tree in fundal images. The Radon transform is less sensitive to noise in the image, because the intensity fluctuations due to noise tend to be cancelled out by the process of integration [39] The Radon transform is able to transform line-containing images into a domain where each line in the image gives a peak or a valley in that domain. In spite of morphologic operators which are image based and couldn’t provide numerical data directly, by using radon transform object finding and description will be done simultaneously. For line detection some numerical data such as line width, and line angle could be achieved directly from Radon matrix.
In the proposed algorithm, Green (G) component of sub-images are selected for applying local Radon transform (RT). The reason is the high contrast between blood vessels and image background in G component. G component is then complemented; this makes the vessels to be associated with peaks in Radon space (not valleys). In spite of the high contrast, we enhance the quality of sub-images before applying RT. Image enhancement improves the sub-image contrast and causes more discrimination between sub-vessels and retinal background and increases the total accuracy. The RT of a function f(x,y), denoted as g(s,θ\theta), is defined as its line integral projected at an angle θ\theta from the y-axis, at a distance s from the origin. The RT, which is a parameter transform, is defined as:

gθ(s)=0y0xf(x,y)δ(sxcosθysinθ)𝑑x𝑑yg_{\theta}(s)=\int_{0}^{y}\int_{0}^{x}f(x,y)\delta(s-xcos\theta-ysin\theta)dxdy (2)

where
- g(s,θ\theta) is the one-dimensional projection of f(x,y) at an angle θ\theta.
- f(x,y) is the image intensity at (x,y).
- δ\delta is the Dirac delta function.
- s = x cosθ\theta + y sinθ\theta , is the distance from the origin to the line being integrated.
The peak and the corresponding projection angle are detected in the Radon matrix. The profile in which peak occurs, is a candidate that might contain a sub- vessel. This profile is further analyzed for validation of candidate sub-vessel.

Refer to caption
Figure 3: (a) Original Green channel fundus image from local database (MUMS-DB), (b) result of segmentation using RT.
Refer to caption
Figure 4: (a) Original fundus image from DRIVE, (b) result of segmentation using RT.

II-C3 Fovea Detection

It is apparent in retinal images that the main vessels diverge from the ONH and follow a specific course that can often be geometrically modeled as parabola, with a common vertex inside the ONH and the fovea is apex of this parabola curve. Based on this assumption, the main retinal blood vessels are segmented and fitted with a parabolic model. We can thus detect the fovea in the fundus images with respect to the entire vascular structure. Here we assume a Cartesian coordinate system, so this parabolic curve can be described as [40]:

x={(x,y):x=ay2}x=\{(x,y):x=ay^{2}\} (3)

Where a is the focus or the amount and direction of opening of the parabola.

Refer to caption
Figure 5: (a) Original Green channel fundus image from local database (MUMS-DB), (b) result of segmentation using RT, (c) result of Fovea detection.
Refer to caption
Figure 6: (a) Original Green channel fundus image from DRIVE (b) result of segmentation using RT, (c) result of Fovea detection.

III Experimental Results

To calculate the efficiency of the current method in fovea detection and also to compare the results with other reported studies, it is necessary to compare all pixels of the final automated segmentated images with the manual segmentation or grand truth (GT) files. For the evaluation, we used the concept of sensitivity and specificity. The results for the automated method compared to the GS were calculated for each image. The higher the sensitivity and specificity values, the better the procedure. Regarding these calculations the proper metrics are defined as [34]:

{Sensitivity=TPTP+FNSpecificity=TNTN+FP\begin{cases}Sensitivity=\frac{TP}{TP+FN}\\ Specificity=\frac{TN}{TN+FP}\\ \end{cases} (4)

Where TP is true positive, TN is true negative, FP is false positive and FN is false negative.

III-A Training and Test Set for the Image Database

In this study, we used 40 images for a training set which was for learning purpose and to set our parameters in the algorithm. This consisted of 20 images from all databases (MUMS-DB, DRIVE). The test set for the test purposes and to evaluate the method consisted of 220 fundus images of which 200 for MUMS-DB, and 20 for the DRIVE. After setting up the parameters of our algorithm using training set, the methods were tested in each image of the databases in test set. Some results are shown in Fig. 5, Fig. 6, and Fig. 7 from MUMS-DB and DRIVE.

Refer to caption
Figure 7: (a) Original Green channel fundus image (b) result of segmentation using RT, (c) result of Fovea detection.

The sensitivity of the threshold was also characterized along the Equation (8). All of three databases, the value ThTh is varied over the range [0, 5]. Parameters used in the algorithm are shown in Table II.

TABLE II: PARAMETERS USED IN OUR APPROACH FOR ALL THE THREE DATABASES
Database No. of Images Window Size (n) Step Th
MUMS-DB 100 62 5 [0,5]
DRIVE 20 15 6 [0,5]
MESSIDOR 100 40 5 [0,5]

Statistical information about the sensitivity and specificity measures is extracted. The higher the sensitivity and specificity values, the better the procedure. For all retinal images of test set (220 images), our reader labeled the fovea on the images and the result of this manual detections were saved to be analyzed further.

The result show that, among 40 images of the DRIVE database we detected fovea in 85% of them. For the MUMS-DB database among 200 images we detect fovea correctly in 83% of them.

IV Discussion and Conclusion

Since fundus images are nowadays in the digital format, it is possible to create a computer-based system that automatically detects landmarks from fundus images [41, 42, 43]. An automatic system would save the working time of well-paid clinicians, and letting hospitals to use their valuable resources in other important tasks. It could also be possible to check more people and more often with the help of an automatic screening system, since it would be more inexpensive than screening by humans. In this study, We present an automatic method for detecting the fovea, that is used for estimating the diabetic macular edema severity levels within the context of this project. Our method identifies the fovea center as a point (a pixel) and has been validated based on two public databases of retinal images. The accuracy of the fovea detection refers to how close is the automatically detected fovea center point (i.e. the distance in pixels) to the fovea center hand-labeled by an expert. Our approach uses a vessel segmentation and reconstruction technique [36] that relies on morphological operations and Radon Transform and a characteristic trait of vessels to predict the location of the fovea. Unlike other works, we make no template from a training set and only use one general feature, parabolic fitting, to detect the fovea. There are however current limitations. We first require a more robust fitting model, one that will need to rotate the image in the instances that the vessels arc neither purely up nor perfectly sideways. Another study needs to be performed to ascertain whether parabolic or higher order fitting is relevant to different or diseased retinas. The second limitation is to improve on the generalization of the location of the fovea when not directly lying on the vertex of the parabola. Future work is to create a line from the apex of the parabola through the vertex and find a region of possible candidate locations for the fovea. This would provide a more comprehensive way to find and detect the fovea. Our proposed algorithm combines many of the powerful techniques from previous works and would be a more generic computational approach that doesn’t require previous exemplary images to base a template on.

One of the most important parts of in this study is to use multi-overlapping window to prevent error caused by global techniques.
One advantage of the automatic screening system is that it is deterministic, in other words, it always classifies funduses in the similar way. There will always exist differences in the backgrounds and education of human screeners, which causes dispersion in their diagnose making. Also a single human expert may make different diagnoses in different screening times due to human factors, such as tiredness or sickness. A computer-based screening system does not have to be perfect to be used in screening. It is better to use a computer-based screening system for classifying only clearly normal funduses as normal, whereas abnormal and obscure funduses are delivered to a human expert for further classification. However, the computer-based screening system reduces the workload of the human expert, since in the screening most of the funduses are normal, and only a few funduses have retinopathy.
Diabetes caused pathological alterations in the retinal blood vessel network that need to detection and treatment as soon as possible in order to prevent DR. For improving the evaluation of the retina condition image processing techniques are required to investigate appropriate data about changes in retinal vasculature. We divided our images into a training set with 40 images and test set with 220 images. The training set was used for developing the algorithms and the test set only for testing the algorithms. Our results for fovea detection are acceptable. The results proved that it is possible to use algorithms for assisting an ophthalmologist to segment fundus images into normal parts and lesions, and thus support the ophthalmologist in his or her decision making. The algorithms detect regions where the image quality is inadequate, and thus it is possible to show to the ophthalmologist what regions are left unprocessed. In the screening it would be important that the computer-based system has very high sensitivity.
The quality of our segmentation depends on some parameters such as the length of our window (n), measure of step, line validation thresholding, etc. determining of these parameter appropriately has some advantages in our processing likes:
Accurate detection of retinal vessels location, Determination of some parameters like width and length of vessels, and Even determining of location of vessel bifurcation which can be assist to clinician for analyzing image in later by registration scheme.
Our algorithm has some important characteristics in detection of vascular structure in retinal images that include:
1. This algorithm is able to determine location, width and angle of vessels.
2. The algorithm is robust to noise.
3. Because of combination of two methods vessel segmentaion methods with the multi-overlapping window the performance of algorithm in detection of thick and even thin vessels is acceptable.

References

  • [1] Khansari MM, Wanek J, Felder AE, Camardo N, Shahidi M. Automated assessment of hemodynamics in the conjunctival microvasculature network. IEEE Transactions on Medical Imaging. 2016 Feb;35(2):605-11.
  • [2] Tavakoli M, Najib M, Abdollahic A, Kalantarid F. Attenuation Correction in SPECT Images Using Attenuation Map Estimation with Its Emission Data. InSPIE Medical Imaging 2017 Mar 9 (pp. 101324Z-101324Z). International Society for Optics and Photonics.
  • [3] Tavakoli M, Mehdizadeh A, Pourreza R, Banaee T, Bahreyni Toossi MH, Pourreza HR. Early Detection of Diabetic Retinopathy in Fluorescent Angiography Retinal Images Using Image Processing Methods. Iranian Journal of Medical Physics. 2010;7(4):7-14.
  • [4] Khansari MM, O’Neill W, Lim J, Shahidi M. Method for quantitative assessment of retinal vessel tortuosity in optical coherence tomography angiography applied to sickle cell retinopathy. Biomedical Optics Express. 2017 Aug 1;8(8):3796-806.
  • [5] Englmeier KH, Schmid K, Hildebrand C, Bichler S, Porta M, Maurino M, Bek T. Early detection of diabetes retinopathy by new algorithms for automatic recognition of vascular changes. European journal of medical research. 2004 Oct 29;9(10):473-8.
  • [6] Hoover AD, Kouznetsova V, Goldbaum M. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Transactions on Medical imaging. 2000 Mar;19(3):203-10.
  • [7] Xu ZW, Guo XX, Hu XY, Cheng X. The blood vessel recognition of ocular fundus. InMachine Learning and Cybernetics, 2005. Proceedings of 2005 International Conference on 2005 Aug 18 (Vol. 7, pp. 4493-4498). IEEE.
  • [8] Cornforth DJ, Jelinek HJ, Leandro JJ, Soares JV, Cesar Jr RM, Cree MJ, Mitchell P, Bossomaier T. Development of retinal blood vessel segmentation methodology using wavelet transforms for assessment of diabetic retinopathy. InThe 8th Asia Pacific Symposium on Intelligent and Evolutionary Systems 2005 (Vol. 11, pp. 50-61).
  • [9] Sinthanayothin C, Boyce JF, Williamson TH, Cook HL, Mensah E, Lal S, Usher D. Automated detection of diabetic retinopathy on digital fundus images. Diabetic medicine. 2002 Feb 1;19(2):105-12.
  • [10] Heneghan C, Flynn J, O’Keefe M, Cahill M. Characterization of changes in blood vessel width and tortuosity in retinopathy of prematurity using image analysis. Medical image analysis. 2002 Dec 31;6(4):407-29.
  • [11] Hoover A, Goldbaum M. Locating the optic nerve in a retinal image using the fuzzy convergence of the blood vessels. IEEE transactions on medical imaging. 2003 Aug;22(8):951-8.
  • [12] Welfer D, Scharcanski J, Marinho DR. Fovea center detection based on the retina anatomy and mathematical morphology. Computer methods and programs in biomedicine. 2011 Dec 31;104(3):397-409.
  • [13] Pourreza HR, Bahreyni Toossi MH, Mehdizadeh A, Pourreza R, Tavakoli M. Automatic Detection of Microaneurysms in Color Fundus Images using a Local Radon Transform Method. Iranian Journal of Medical Physics. 2009 Mar 1;6(1):13-20.
  • [14] Niemeijer M, Abràmoff MD, Van Ginneken B. Fast detection of the optic disc and fovea in color fundus photographs. Medical image analysis. 2009 Dec 31;13(6):859-70.
  • [15] Sinthanayothin C, Boyce JF, Cook HL, Williamson TH. Automated localisation of the optic disc, fovea, and retinal blood vessels from digital colour fundus images. British Journal of Ophthalmology. 1999 Aug 1;83(8):902-10.
  • [16] Gagnon L, Lalonde M, Beaulieu M, Boucher MC. Procedure to detect anatomical structures in optical fundus images. InProc. SPIE 2001 Feb 19 (Vol. 4322, pp. 1218-1225).
  • [17] Sherwani SM, Tiwana MI, Iqbal J, Lovell NH. Automated Segmentation of Optic Disc Boundary and Diameter Calculation Using Fundus Imagery. InProceedings of the 2015 Seventh International Conference on Computational Intelligence, Modelling and Simulation 2015 Jul 27 (pp. 92-96). IEEE Computer Society.
  • [18] Muangnak N, Aimmanee P, Makhanov S, Uyyanonvara B. Vessel transform for automatic optic disk detection in retinal images. IET Image Processing. 2015 Sep 1;9(9):743-50.
  • [19] Constante P, Gordón A, Chang O, Pruna E, Escobar I. Neural networks for optic nerve detection in digital optic fundus images. InAutomatica (ICA-ACCA), IEEE International Conference on 2016 Oct 19 (pp. 1-5). IEEE.
  • [20] Khansari MM, O’Neill W, Penn R, Blair NP, Chau F, Shahidi M. An automated image processing method for classification of diabetic retinopathy stages from conjunctival microvasculature images. InSPIE Medical Imaging 2017 Mar 13 (pp. 101372C-101372C). International Society for Optics and Photonics.
  • [21] Staal J, Abràmoff MD, Niemeijer M, Viergever MA, Van Ginneken B. Ridge-based vessel segmentation in color images of the retina. IEEE transactions on medical imaging. 2004 Apr;23(4):501-9.
  • [22] Soares JV, Leandro JJ, Cesar RM, Jelinek HF, Cree MJ. Retinal vessel segmentation using the 2-D Gabor wavelet and supervised classification. IEEE Transactions on medical Imaging. 2006 Sep;25(9):1214-22.
  • [23] Jiang X, Mojon D. Adaptive local thresholding by verification-based multithreshold probing with application to vessel detection in retinal images. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2003 Jan;25(1):131-7.
  • [24] Wu D, Zhang M, Liu JC, Bauman W. On the adaptive detection of blood vessels in retinal images. IEEE Transactions on Biomedical Engineering. 2006 Feb;53(2):341-3.
  • [25] Tolias YA, Panas SM. A fuzzy vessel tracking algorithm for retinal images based on fuzzy clustering. IEEE Transactions on Medical Imaging. 1998 Apr;17(2):263-73.
  • [26] Zana F, Klein JC. Segmentation of vessel-like patterns using mathematical morphology and curvature evaluation. IEEE transactions on image processing. 2001 Jul;10(7):1010-9.
  • [27] Walter T, Klein JC. Segmentation of color fundus images of the human retina: Detection of the optic disc and the vascular tree using morphological techniques. Medical data analysis. 2001:282-7.
  • [28] Tavakoli M, Toosi MB, Pourreza R, Banaee T, Pourreza HR. Automated optic nerve head detection in fluorescein angiography fundus images. InNuclear Science Symposium and Medical Imaging Conference (NSS/MIC), 2011 IEEE 2011 Oct 23 (pp. 3057-3060). IEEE.
  • [29] Singh J, Joshi GD, Sivaswamy J. Appearance-based object detection in colour retinal images. InImage Processing, 2008. ICIP 2008. 15th IEEE International Conference on 2008 Oct 12 (pp. 1432-1435). IEEE.
  • [30] Sekhar S, Al-Nuaimy W, Nandi AK. Automated localisation of optic disk and fovea in retinal fundus images. InSignal Processing Conference, 2008 16th European 2008 Aug 25 (pp. 1-5). IEEE.
  • [31] Fleming AD, Goatman KA, Philip S, Olson JA, Sharp PF. Automatic detection of retinal anatomy to assist diabetic retinopathy screening. Physics in medicine and biology. 2006 Dec 21;52(2):331.
  • [32] Li H, Chutatape O. Automated feature extraction in color retinal images by a model based approach. IEEE Transactions on biomedical engineering. 2004 Feb;51(2):246-54.
  • [33] Tobin KW, Chaum E, Govindasamy VP, Karnowski TP. Detection of anatomic structures in human retinal imagery. IEEE transactions on medical imaging. 2007 Dec;26(12):1729-39.
  • [34] Tavakoli M, Shahri RP, Pourreza H, Mehdizadeh A, Banaee T, Toosi MH. A complementary method for automated detection of microaneurysms in fluorescein angiography fundus images to assess diabetic retinopathy. Pattern Recognition. 2013 Oct 31;46(10):2740-53.
  • [35] Pourreza-Shahri R, Tavakoli M, Kehtarnavaz N. Computationally efficient optic nerve head detection in retinal fundus images. Biomedical Signal Processing and Control. 2014 May 31;11:63-73.
  • [36] Tavakoli M, Mehdizadeh AR, Pourreza R, Pourreza HR, Banaee T, Toosi MB. Radon transform technique for linear structures detection: application to vessel detection in fluorescein angiography fundus images. InNuclear Science Symposium and Medical Imaging Conference (NSS/MIC), 2011 IEEE 2011 Oct 23 (pp. 3051-3056). IEEE.
  • [37] Research Section, Digital Retinal Image for Vessel Extraction (DRIVE) Database. Utrecht, The Netherlands, Univ. Med. Center Utrecht, Image Sci. Inst. [Online]. Available: http://www.isi.uu.nl/Re-search/Databases/DRIVE
  • [38] Tavakoli M, Nazar M, Mehdizadeh A. Effect of Two Different Preprocessing Steps in Detection of Optic Nerve Head in Fundus Images. In SPIE Medical Imaging 2017 Mar 3 (pp. 101343A-101343A). International Society for Optics and Photonics.
  • [39] Choi DI, Park SH. Self-creating and organizing neural networks. IEEE Transactions on Neural Networks. 1994 Jul;5(4):561-75.
  • [40] Foracchia M, Grisan E, Ruggeri A. Detection of optic disc in retinal images by means of a geometrical model of vessel structure. IEEE transactions on medical imaging. 2004 Oct;23(10):1189-95.
  • [41] Welikala RA, Dehmeshki J, Hoppe A, Tah V, Mann S, Williamson TH, Barman SA. Automated detection of proliferative diabetic retinopathy using a modified line operator and dual classification. Computer Methods and Programs in Biomedicine. 2014 May 31;114(3):247-61.
  • [42] Welikala RA, Fraz MM, Dehmeshki J, Hoppe A, Tah V, Mann S, Williamson TH, Barman SA. Genetic algorithm based feature selection combined with dual classification for the automated detection of proliferative diabetic retinopathy. Computerized Medical Imaging and Graphics. 2015 Jul 31;43:64-77.
  • [43] Khansari MM, O’Neill W, Penn R, Chau F, Blair NP, Shahidi M. Automated fine structure image analysis method for discrimination of diabetic retinopathy stage using conjunctival microvasculature images. Biomedical optics express. 2016 Jul 1;7(7):2597-606.