Abstract
We present a new framework for fine-scale vessel segmentation from fundus images through registration and segmentation of corresponding fluorescein angiography (FA) images. In FA, fluorescent dye is used to highlight the vessels and increase their contrast. Since these highlights are temporally dispersed among multiple FA frames, we first register the FA frames and aggregate the per-frame segmentations to construct a detailed vessel mask. The constructed FA vessel mask is then registered to the fundus image based on an initial fundus vessel mask. Postprocessing is performed to refine the final vessel mask. Registration of FA frames, as well as registration of FA vessel mask to the fundus image, are done by similar hierarchical coarse-to-fine frameworks, both comprising rigid and non-rigid registration. Two CNNs with identical network structures, both trained on public datasets but with different settings, are used for vessel segmentation. The resulting final vessel segmentation contains fine-scale, filamentary vessels extracted from FA and corresponding to the fundus image. We provide quantitative evaluation as well as qualitative examples which support the robustness and the accuracy of the proposed method.
This work was supported by the National Research Foundation of Korea (NRF) grants funded by the Korean government (MoE) (NRF-2018R1D1A1A09083241 and NRF-2019R1F1A1063656).
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Retinal fundus images are the only type of medical image that directly observes the blood vessels to generate clear, high resolution visualizations. They are simple, noninvasive, relatively cheap, and require no radiation or pharmaceuticals. They are used to diagnose various retinal diseases including diabetic retinopathy, age-related macular degeneration, epiretinal membrane, and glaucoma. They can also be used for early diagnosis and prevention of chronic diseases including diabetes and hypertension.
Chronic diseases can damage vessels and also cause new vessels to be formed [15]. Clinicians require highly accurate detection and measurement of vessels, including fine, filamentary vessels with thin complex shapes and low contrast, for better diagnoses. Thus, this problem has been extensively researched [1].
Public retinal image datasets including DRIVE [14], STARE [9], CHASE_DB1 [6], and HRF [3] have been vital to the research. These datasets all include vessel region masks achieved by manual expert annotations. While they are assumed as the ground truth, it is actually very difficult to measure their accuracy. Comparison with a second expert actually show the limitations of human annotations. The accuracy of recent automatic vessel segmentation methods [11, 13] are higher than the annotations by the second expert. This is because inter-observer differences inevitably occur at ambiguous regions. Filamentary vessels are often barely visible in the retinal fundus image, even with zooming and contrast enhancement. If we can provide consistent and detailed annotations at these regions, ground truth expert annotations can be improved, which in turn can improve machine learning based methods.
In this paper we present a new framework for retinal vessel extraction from fundus images through registration and segmentation of corresponding fluorescein angiography (FA) images. In FA, a fluorescent dye is injected into the bloodstream to highlight the vessels by increasing contrast. But the highlights are temporally dispersed among the multiple FA frames while the dye flows through the vessels from arterioles to venules. Thus we must first align the FA frames and aggregate the per-frame segmentations to construct a detailed vessel mask. Here, alignment is done by keypoint based registration, and vessel segmentation is done using a convolutional neural net (CNN) [11]. The constructed FA vessel mask is then registered to the fundus image based on an initial vessel segmentation for the fundus image, again using a CNN. Postprocessing is performed to refine the final fundas image vessel mask based on the FA vessel mask. We believe the proposed method is the first method to successfully elevate the level of detail and accuracy in automatic vessel segmentation of filamentary vessels for fundus images by incorporating FA. Moreover, it can be used to generate more detailed and consistent ground truth vessel masks as shown in Fig. 1.
2 Methods
The proposed method can be compartmentalized into three subprocesses, corresponding to (1) registration of FA frames and their vessel extraction, (2) multi-modal registration of aggregated FA vessels to the fundus image, (3) postprocessing for fine refinement of the vessel mask. We describe the details of each subprocess in the following subsections.
2.1 Registration and Vessel Extraction of FA Frames
Here, the objective is to extract a mask of all vessels including the filamentary ones. We thus extract all vessels from all FA frames and aggregate them in a combined registered frame. In contrast to methods based on registration of extracted vessels [12], vessels are extracted after registration, since the highlighted vessels change considerably due to blood flow. Moreover, registration actually helps the vessel extraction, since false positives can be avoided through the aggregation of vessel regions. Thus, we propose a three step hierarchical process, combining coarse rigid registration in the pixel domain and fine non-rigid registration in the vessel mask domain to ensure robustness against appearance changes of the frames and its vessels. A visual summary of the framework is given in Fig. 2. We note that this process is iteratively performed for all adjacent frame pairs with the initial frame as the reference frame.
In the first step, feature point matching is performed in the pixel domain. We use the SIFT descriptor on local maxima in the difference of Gaussians [10]. Keypoint matching is performed using RANSAC (random sample consensus) with the perspective transform matrix [8]. The source image is then rigidly registered to the target image using the transform matrix determined by the keypoint matches.
We next refine the rigid registration by non-rigid registration of vessel probability maps. Here, we leverage recent developments in deep learning by using a recently proposed convolutional neural network (CNN) called the retinal SSANet [11]. Since no training data is available for supervised learning of the network for the FA frames, we utilize the public datasets DRIVE [14] and HRF [3] comprising fundus images and expert annotated ground truth vessel maps. To account for the differences in image characteristics, we convert to greyscale and then invert intensity before training. We also resize the images to match FA resolution.
Given the vessel probability maps, we then perform pixel-wise non-rigid registration. We assume a b-spline transform model with similarity measured by normalized cross-correlation and optimization with the gradient based L-BFGS-B [4] algorithm.
2.2 Registration of FA and Fundus Image
To register the aggregated probabilistic vessel map of FA, we generate a similar map for the fundus image. Again, we train a retinal SSANet [11], this time without any preprocessing on images of DRIVE and HRF images. The vessel maps are generated from inference of this network. Based on the vessel maps, we again perform coarse rigid registration, this time using chamfer matching, followed by fine non-rigid registration. For chamfer matching, we first assign the binarized fundus image and FA vessel masks as the source and target shapes. We then find the global displacement vector and the rotation angle (within a \(\pm 5 ^\circ \) range) that minimizes the sum of distances between each point on the source shape and the target shape by brute force search on the distance transform (DT) of the target shape. We use the inverse of the obtained transform to align the FA map to the fundus image. For non-rigid registration, we use the same specifics as in Subsect. 2.1. A visual summary of the framework is given in Fig. 3
2.3 Postprocessing
Here, we aim to generate an accurate binary vessel mask of the fundus image, from the aligned probabilistic vessel map of the FA. The postprocessing comprises binarization and refinement. A visual summary is given in Fig. 4.
To avoid discontinuities that may occur at the filamentary vessels from simple thresholding, we apply hysteresis thresholding for binarization. Pixels over a higher threshold \(\tau _{h}\) are used as seeds for region growing pixels with probability higher than a lower threshold \(\tau _{l}\). Here, we empirically set \(\tau _{h}=0.75\) and \(\tau _{l}=0.1\).
Furthermore, we refine the vessel mask to align the vessel boundaries to the image gradients in the fundus image. Specifically, we utilize the Frangi filter [5], in an inverted manner with sigma values 1 to 3, to detect the valleys between vessels and the outer boundaries, and then erode these regions.
3 Experimental Results
3.1 Dataset and Experimental Environment
The dataset comprises 200 cases of FA and fundus image pairs from 153 patients. They were acquired using Canon CF60Uvi, Kowa VX-10, and Kowa VX-10a cameras. The number of FA frames was on average 7.14, with minimum 2 to maximum 24. Image resolutions originally varying from \(1604 \times 1216\) to \(2144 \times 1424\) were all normalized to \(1536 \times 1024\).
Computation times for the FA registration per frame pair, FA-fundus registration, and postprocessing averaged 57, 28, and 3 s, respectively, running on a 2.2 GHz Intel Xeon CPU and a nVidia Titan V GPU. Most of the computation was due to feature point matching (over 40 of the 57 s), and non-rigid registration took 15–16 s on average. OpenCV [2] was used for feature point matching and SimpleITK [16] was used for non-rigid registration.
3.2 Qualitative Evaluation
Figure 5 shows qualitative results of six sample cases. Here, we provide the vessel segmentation results generated by a CNN, namely, the SSANet [11] trained on the HRF [3] dataset, as a reference point for comparison. Although we are aware that this comparison maybe unfair, we were unable to establish an alternative comparative reference. Here, we can see that many filamentary vessels are indeed visible in the fundus images, but only with close visual inspection. Figure 6 shows a particular example of this case.
3.3 Quantitative Evaluation
Ground truth (GT) segmentation masks are required for quantitative evaluation. But we cannot rely on expert annotation for filamentary vessels. We thus generate GT masks by manually editing the results from the proposed method. Editing mostly comprised removal of false positives near the optic disk by direct annotation. The average duration was 53 s per image.
We compared the results of the proposed method with the aforementioned SSANet [11] trained with different public training datasets including DRIVE [14], STARE [9], CHASE_DB1 [6], and HRF [3], in Table 1. The networks were trained on the resolution of the images in each dataset, and fundus images of our FA-fundus image set were resized accordingly and given as input images. Measures were computed based on the aforementioned GT. We present these results as a reference for understanding the performance of the proposed method.
4 Discussion
We present a new method to generate fine-scale vessel segmentation masks for fundus images by registration with FA. We have shown that the obtained results contain a considerable amount of filamentary vessels that are virtually indiscernible to the naked eye with only the color fundus image. We believe that these results conversely show the limitations of expert annotations as ground truth, which is the standard of all previously released public datasets. Nonetheless, since the proposed method may still contain errors, the requirement of expert annotation remains in order to designate data as ground truth.
For future works, we plan to establish better means of quantitative evaluations for the proposed method. We are aware of the biases of GT toward the proposed method in our current quantitative evaluations. Unfortunately, methods such as that of Galdran et al. [7], that learns to estimate accuracy from existing GT, is inapplicable since it relies on existing expert annotations.
Also, we plan to construct a new dataset for filamentary vessels that can be used for improving deep learning based methods for retinal vessel segmentation. Our ultimate aim is to construct a dataset that can be the foundation for achieving superhuman accuracy. Particularly, although we utilize FA to construct the ground truth for this dataset, our intention is to use the generated ground truth for supervised learning of a vessel segmentation method with only fundus images as input.
References
Almotiri, J., Elleithy, K., Elleithy, A.: Retinal vessels segmentation techniques and algorithms: a survey. Appl. Sci. 8(2), 155 (2018). https://doi.org/10.3390/app8020155
Bradski, G.: The OpenCV library. Dr. Dobb’s J. Softw. Tools 25, 120–125 (2000)
Budai, A., Bock, R., Maier, A., Hornegger, J., Michelson, G.: Robust vessel segmentation in fundus images. Int. J. Biomed. Imaging 2013, 154860 (2013)
Byrd, R.H., Lu, P., Nocedal, J., Zhu, C.: A limited memory algorithm for bound constrained optimization. SIAM J. Sci. Comput. 16(5), 1190–1208 (1995)
Frangi, A.F., Niessen, W.J., Vincken, K.L., Viergever, M.A.: Multiscale vessel enhancement filtering. In: Wells, W.M., Colchester, A., Delp, S. (eds.) MICCAI 1998. LNCS, vol. 1496, pp. 130–137. Springer, Heidelberg (1998). https://doi.org/10.1007/BFb0056195
Fraz, M.M., et al.: An ensemble classification-based approach applied to retinal blood vessel segmentation. IEEE Trans. Biomed. Eng. 59(9), 2538–2548 (2012). https://doi.org/10.1109/TBME.2012.2205687
Galdran, A., Costa, P., Bria, A., Araújo, T., Mendonça, A.M., Campilho, A.: A no-reference quality metric for retinal vessel tree segmentation. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11070, pp. 82–90. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00928-1_10
Hartley, R., Zisserman, A.: Multiple View Geometry in Computer Vision, 2nd edn. Cambridge University Press, New York (2003)
Hoover, A.D., Kouznetsova, V., Goldbaum, M.: Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 19(3), 203–210 (2000). https://doi.org/10.1109/42.845178
Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60(2), 91–110 (2004). https://doi.org/10.1023/B:VISI.0000029664.99615.94
Noh, K.J., Park, S.J., Lee, S.: Scale-space approximated convolutional neural networks for retinal vessel segmentation. Comput. Methods Programs Biomed. 178, 237–246 (2019). https://doi.org/10.1016/j.cmpb.2019.06.030
Perez-Rovira, A., Trucco, E., Wilson, P., Liu, J.: Deformable registration of retinal fluorescein angiogram sequences using vasculature structures. In: International Conference of the IEEE Engineering in Medicine and Biology (EMBS), pp. 4383–4386, August 2010. https://doi.org/10.1109/IEMBS.2010.5627094
Son, J., Park, S.J., Jung, K.H.: Towards accurate segmentation of retinal vessels and the optic disc in fundoscopic images with generative adversarial networks. J. Digit. Imaging 32, 499–512 (2018). https://doi.org/10.1007/s10278-018-0126-3
Staal, J., Abramoff, M.D., Niemeijer, M., Viergever, M.A., van Ginneken, B.: Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 23(4), 501–509 (2004). https://doi.org/10.1109/TMI.2004.825627
Viswanath, K., McGavin, D.D.M.: Diabetic retinopathy: clinical findings and management. Community Eye Health 16(46), 21–24 (2003). https://www.ncbi.nlm.nih.gov/pubmed/17491851
Yaniv, Z., Lowekamp, B.C., Johnson, H.J., Beare, R.: SimpleITK image-analysis notebooks: a collaborative environment for education and reproducible research. J. Digit. Imaging 31(3), 290–303 (2018). https://doi.org/10.1007/s10278-017-0037-8
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Noh, K.J., Park, S.J., Lee, S. (2019). Fine-Scale Vessel Extraction in Fundus Images by Registration with Fluorescein Angiography. In: Shen, D., et al. Medical Image Computing and Computer Assisted Intervention – MICCAI 2019. MICCAI 2019. Lecture Notes in Computer Science(), vol 11764. Springer, Cham. https://doi.org/10.1007/978-3-030-32239-7_86
Download citation
DOI: https://doi.org/10.1007/978-3-030-32239-7_86
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-32238-0
Online ISBN: 978-3-030-32239-7
eBook Packages: Computer ScienceComputer Science (R0)