Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Region-segmentation strategy for Bruch’s membrane opening detection in spectral domain optical coherence tomography images

Open Access Open Access

Abstract

Bruch’s membrane opening (BMO) is an important biomarker in the progression of glaucoma. Bruch’s membrane opening minimum rim width (BMO-MRW), cup-to-disc ratio in spectral domain optical coherence tomography (SD-OCT) and lamina cribrosa depth based on BMO are important measurable parameters for glaucoma diagnosis. The accuracy of measuring these parameters is significantly affected by BMO detection. In this paper, we propose a method for automatically detecting BMO in SD-OCT volumes accurately to reduce the impact of the border tissue and vessel shadows. The method includes three stages: a coarse detection stage composed by retinal pigment epithelium layer segmentation, optic disc segmentation, and multi-modal registration; a fixed detection stage based on the U-net in which BMO detection is transformed into a region segmentation problem and an area bias component is proposed in the loss function; and a post-processing stage based on the consistency of results to remove outliers. Experimental results show that the proposed method outperforms previous methods and achieves a mean error of 42.38 μm.

© 2019 Optical Society of America under the terms of the OSA Open Access Publishing Agreement

1. Introduction

Glaucoma is an irreversible, progressive optic neuropathy that causes nerve fiber damage, visual loss, and even blindness [1, 2]. Early detection and treatment can slow the progression of the disease; thus, identification of the structural changes in early glaucoma, such as cup-to-disc ratio and retinal nerve fiber layer (RNFL) thickness, is important [3, 4].

Bruch’s membrane opening (BMO) or neural canal opening, which remains stable with the progression of glaucoma, is an important biomarker in the optic nerve head (ONH) in Spectral-Domain Optical Coherence Tomography (SD-OCT) [5]. It is defined as the termination of the retinal pigment epithelium (RPE) layer. As introduced in [6], the BMO points in SD-OCT represent the true position of the optic disc. Thus, the disc margin can also be identified by aligning the color fundus image and the SD-OCT volumes.

Bruch’s membrane opening minimum rim width (BMO-MRW), defined as the minimum Euclidean distance from BMO to the internal limiting membrane surface, performs better in early glaucoma diagnosis than other structure parameters, such as RNFL thickness [7]. Moreover, BMO plays an indispensable role in lamina cribrosa surface depth measurement [8], which is also an important parameter for measuring the progression of glaucoma [9]. Hence, an accurate and robust detection method of BMO is needed to measure those structural parameters automatically.

Various methods have been employed to segment the optic disc in color fundus images [10–13], but peripapillary atrophy can exert a great impact when the brightness of the disc region is relatively low. Thus, scholars have proposed some approaches for disc segmentation in SD-OCT images. Hu et al. [14] proposed a graph-based method that segments the RPE layer by 3D graph search. Miri et al. [15] proposed an advanced approach that utilizes the random forest to learn the cost of graph search. Fu et al. [16] proposed a low-rank based method that uses low-rank decomposition to segment the disc in SD-OCT images.

The approaches focused on BMO detection mainly attempted to segment the BMO points in a single OCT B-scan. Beighth et al. [17] proposed a model-based method to identify the RPE layer and BMO and then used the disc shape as a priori to obtain the best curve that represents BMO. The graph-based method proposed in [14] segmented BMO in a single OCT slice. However, it can be influenced by vessel shadows and speckle noise. The three-bench mark reference (TBMR) method proposed by Hussuain et al. [18] used a constraint of the location of three bench mark reference layers to correct the result of graph search. TBMR reduced the impact of the border tissue and vessel shadows and obtains a better result.

Several approaches have also been used to detect BMO in SD-OCT volumes. Wu et al. [19] proposed a patch searching method that transforms BMO detection to searching the image patch that is centered on the BMO points in 3D SD-OCT volumes. The method first applies 3D graph search to accomplish coarse detection, determines the region of interest (ROI), and then divides the patch in the ROI into four classes, namely, patch centered on the left BMO, patch centered on the right BMO, patch in the RPE layer, and background. Finally, the method utilizes a support vector machine to classify the patches and obtain results. However, the results obtained by this method are not robust to the border tissue. Miri et al. [20] proposed an advanced version of the method in [15] to find the 3D BMO path. The method employs a dynamic programming algorithm to find the shortest path in high accuracy and good robustness. Given that the detection accuracy determines the measurement precision of parameters, the aforementioned methods are not precise enough because of the influence of the border tissue and vessel shadows.

Deep learning based method had been used in some close tasks such as the layer segmentation [21] and fluid detection [22] of SD-OCT, however, it has not been applied for BMO detection yet. In this paper, we propose a method to reduce the effects of the border tissue and vessel shadows and then further improve the precision of BMO detection. This method contains a coarse detection stage based on the RPE layer segmentation and disc segmentation in 2D projection images, a fixed detection stage based on U-net, and a post-processing stage to remove the outliers caused by the border tissue. The main contributions of our proposed method are as follows:

  1. The deep convolutional neural network (CNN) is first utilized to solve the problem of BMO detection in optic nerve head SD-OCT volumes to our best knowledge. To deal with the imbalanced classes, we extend the BMO points into a circular region. This extension allows more neighborhood information to be used to reduce the effect of peripapillary atrophy.
  2. To improve the accuracy of BMO detection, we add an area bias in the dice loss function. Thus, improved detection results are obtained because the area priori information is utilized.
  3. A post-processing algorithm after our region-segmentation strategy is proposed to remove the outliers caused by the border tissue and further improve the accuracy.

2. Methods

In this section, our proposed method to decrease the effects of the border tissue and achieve a precise detection is introduced. The overall flowchart of our proposed method is shown in Fig. 1. The three main stages are 1) coarse detection (Section 2.1), 2) fixed detection based on U-net (Section 2.2), and 3) post-processing (Section 2.3). In the first stage, the coarse detection results are produced by employing graph-based layer segmentation and registration with the color fundus image and the 2D projection image, which are created by adding each A-scan in an OCT volume. In the second stage, we transform the detection to a region segmentation problem and utilize the U-net to obtain a fixed detection result. Finally, a post-processing algorithm is proposed to remove the outliers.

 figure: Fig. 1

Fig. 1 Flowchart of the proposed method.

Download Full Size | PDF

2.1. Coarse detection

Considering the difficulty of BMO detection in an entire OCT slice, a coarse detection to determine the ROI is needed. Fig. 2 indicates that the coarse detection contains two steps: optic disc segmentation and RPE layer segmentation.

Optic disc segmentation

As BMO is the true position of the optic disc, segmenting the disc can be a reasonable way to locate BMO. Segmentation of the optic disc in the 2D projection image has three steps. First, each B-scan is denoised by employing the block-matching 3D (BM3D) algorithm [23], and 3D SD-OCT volumes are resampled to create 2D projection images, the 2D projection image is created by averaging the voxel intensities in the z-direction in the SD-OCT volumes, as shown in Fig. 2(a). Second, the optic disc in the color fundus image is segmented [24]. Third, a robust registration algorithm [25], which is shown in Fig. 2(b), is used to align the color fundus image and the projection image. In the registration algorithm, the features from accelerated segment test-partial intensity invariant feature descriptor framework are used to extract the interest points from the multi-modal images, and then a single Gaussian robust point matching model is applied to match the features and eliminate the outliers.

 figure: Fig. 2

Fig. 2 Coarse detection. (a) Original SD-OCT images and 2D projection images. (b) Registration between the projection and fundus images. (c) Diagram of the coarse detection, intersection (yellow) (yellow) of the RPE layer (red), and projection line of the disc (green) are the coarse detection results.

Download Full Size | PDF

RPE layer segmentation

A conditional random field (CRF) based method [26] is utilized to segment the RPE layer. The method trains a structured support vector machine to construct the energy of the CRF to segment the layers, the coarse BMO is obtained according to the result of disc segmentation and RPE layer segmentation. Moreover, the RPE layer segmentation also used to deal with the B-scan that near the boundary of the segmented disc. In this case, we segment the RPE layer of five B-scans out of the topmost and bottom points of the coarse optic disc boundary and get the inflection point as the coarse BMO in order to contain as many images with BMO as possible.

As shown in Fig. 2(c), the coarse detection result is the intersection of the RPE layer and the projection line of the optic disc. Once the results of coarse detection are obtained, the ROI is selected as an 80×80 patch centered on the coarse detection points, as shown in Fig. 3(a).

2.2. Fixed detection based on U-net

Given the influence of optic disc segmentation and registration, the coarse detection results show relatively low accuracy and stability. In this section, we introduce a fixed detection stage based on the CNN to further improve the accuracy.

2.2.1. Data process for training

In the training stage, the two ROIs (one represents the left BMO and the other represents the right) in each OCT B-scan are selected according to the ground truth. We randomly obtain several 64×64 patches in each ROI region. To further enlarge the dataset, we utilize the augmentation containing random rotation and horizontal flips.

Considering the difficulty for the model to detect a single point in the patch, we extend the label to a circle region centered at BMO with a radius r and then transform the problem to the segmentation task of the circular region. As a result, the effect of imbalanced data is decreased and additional neighborhood information is learned to distinguish BMO and the border tissue. Fig. 3(b) shows the labels that are transformed into the circular region.

 figure: Fig. 3

Fig. 3 Data process.(a) ROI determination. (b) Label transformation for different patches.

Download Full Size | PDF

2.2.2. Network architecture and training details

The U-net [27] is efficient in the domain of medical image segmentation. Thus, the fixed detection stage in our proposed method is based on U-net, as shown in Fig. 4.

 figure: Fig. 4

Fig. 4 U-net architecture in our proposed method.

Download Full Size | PDF

Considering the unavoidable noises and border tissues around BMO, we combine the denoised image generated by BM3D with the original image to compose a dual-channel image. Then, the dual-channel image is sent to U-net to produce improved detection results.To avoid over-fitting, we employ dropout and L2 normalization.

We use the Stochastic Gradient Descent optimizer with 0.9 momentum. The learning rate is set to 0.001 and is gradually decreased. The dropout rate is 0.5, and the λ of L2 normalization is 1e-5.

2.2.3. Loss function

Given the remaining problem of the imbalanced class, the loss function in our network is based on the dice loss. The dice coefficient is a general measure for segmentation, first proposed by Milletari et al. [28] as a loss function for the medical image analysis. The dice loss is formulated as Eq. (1):

DL=12xΩpxgxxΩpx2xΩgx2
where px and gx represent the prediction result and ground truth with the pixel x in the patch Ω, respectively, xΩpx and xΩgx represent the number of pixels in prediction and ground truth region, so that sumxΩpxgx indicates the true positive area of prediction compared to the ground truth.

As mentioned above, the circular region centered on the true BMO is marked as the ground truth. However, the area of the output is always too small to achieve a high dice coefficient. To solve this problem, we add an area bias in the dice loss, formulated as Eq. (2):

AL=1min(xΩpx,xΩgx)max(xΩpx,xΩgx)

To balance the accuracy of the output, we add the mean square error (MSE) loss because it is smoother than the cross-entropy. The final loss function of our proposed method is given by Eq. (3):

Loss=DL+AL+MSE =22xΩpxgxxΩpx2xΩgx2min(xΩpx,xΩgx)max(xΩpx,xΩgx)+1nxΩ(pxgx)2
where n is the summary of the pixels in the region Ω.

 figure: Fig. 5

Fig. 5 Illustration of the border tissue and the post-processing steps. (a) The sketch map of ambiguous border tissue. The real BMO is confused by the border tissue in this case. (b) The sketch map of post-processing. There are four patches (the box in the ROI indicate the upper left patch) extracted in the ROI and send to the trained U-net, the mask of patches with maximum Mi is selected as the result and its center point is the final result of BMO.

Download Full Size | PDF

2.3. Post-processing

The border tissue caused by peripapillary atrophy is highly similar to BMO [29], as shown in Fig. 5 (a), which confuses the model to identify the BMO region centered on the border tissue instead of the end of the BM surface and causes some outliers. To remove the outliers caused by the border tissue, we propose a post-processing algorithm. Considering that most of the results of patches produced by the network are correct, we can utilize the correct results and eliminate the outliers. The ROI of the test image is cropped into several patches in an overlapping manner, and then several patches are sent to the trained network simultaneously. Finally, we choose the best result by the metrics defined in Eq. (4):

Mi=j=1,ijmDice(Si,Sj)k
where m is the patch of each test image. For adapting the size of ROI and patches, Si,Sj indicates the segmentation map of the i-th and j-th patches of the ROI image I, Dice(Si,Sj)=xΩSixSjxxΩSix+xΩSjx is the dice coefficient of the Si,Sj, and k is an exponential parameter which is used to enlarge the consistent components and eliminate the effects of small dice value. At last, the center of the segmentation map is extracted as the final results of BMO detection, formulated as Eq. (5):
C=Center(Sd),d=arg max iMi,i=1,2m
where d is the index of the patch that has a maximum M, Center() represents the geometric center of the segmentation map, and C indicates the results of BMO points. The sketch map of post-processing is shown as Fig. 5(b).

3. Results

3.1. Dataset

In this work, our proposed method is evaluated in 30 SD-OCT volumes, which are collected from the device Topcon 3D OCT-1 Maestro at the second Xiangya Hospital of Central South University. The size of each volume is 885×512×128 voxels, which represents 6 mm×6 mm×2 mm in the x-y-z direction. We use 80% and 20% of the dataset to train and test the U-net, respectively. The BMO points are marked by experts on all B-scans centered on the ONH region.

3.2. Experiments on method design

In this section, we describe a set of experiments to study the impact of different parameters, including radius parameters of the circular region, loss function components, and post-processing, on the performance of our proposed method as assessed by computing the accuracy of region segmentation and the mean error of BMO detection.

3.2.1. Effect of radius parameters of the circular region

The radius of the circular region indicates the neighborhood information sent to the U-net. For instance, the imbalanced class cannot be effectively solved if the radius is overly small, and unnecessary background could mislead the model if the radius is overly large.

In order to exclude the effects of area bias, we evaluated various radius parameters on the model with and without area bias. The results obtained with different radii are shown in Fig. 6. Regardless of the radius, the mean error becomes significantly larger than the properly radius, especially in the model results with loss function without area bias. We choose the radius = 9 in our experiment.

 figure: Fig. 6

Fig. 6 Results of various radius parameters.

Download Full Size | PDF

3.2.2. Effect of loss function components

The effectiveness of our proposed loss function is evaluated through comparisons with simplified versions of our proposed loss function with only a subset of the three components mentioned above: 1) dice, 2) dice+MSE, 3) dice+area bias, and 4) dice+area bias+MSE.

As shown in Table 1, the combination of dice loss, MSE, and our proposed area bias outperforms the other subsets. Adding the proposed area bias to the loss function improves the result to 42.38 μm in the mean error, which indicates the benefit of the area bias. In addition, the MSE loss components further improve the performance by preventing the network from predicting two separate small regions caused by the vessel shadows.

Tables Icon

Table 1. Evaluation of the effect of loss function components

Fig. 7(a) shows the comparison of groups 2 and 4. The result of loss function without the area bias can be influenced by the border tissue, but group 4 avoids this problem and is highly consistent to the ground truth. Fig. 7(b) shows that the result accuracy decreases in the absence of MSE components.

 figure: Fig. 7

Fig. 7 Result of various loss function components. (a) Results of group 4 (green), group 2 (yellow), and the ground truth (red). (b) Results of group 4 (green), group 3 (pink), and the ground truth (red). Best viewed in color.

Download Full Size | PDF

3.2.3. Effect of the post-processing

To evaluate our proposed post-processing algorithm, we compared the method without post-processing with our proposed method under several conditions. As shown in Table 2, the mean error of the method without post-processing is 5-7 μm larger than that of the method with post-processing, indicating that our proposed post-processing algorithm plays a significant role in our detection method. Furthermore, the result of the method without post-processing proves the benefits of area bias loss components.

Tables Icon

Table 2. Evaluation of the effect of post-processing

 figure: Fig. 8

Fig. 8 Evaluation of the effect of post-processing. The dots in red, green, and yellow represent the ground truth, the result with post-processing, and the result without post-processing, respectively. (a) Result of the method without post-processing is satisfactory. (b) Result of the method without post-processing has a large deviation. Best viewed in color.

Download Full Size | PDF

As shown in Fig. 8(a), the results of the method with and without post-processing and the ground truth are almost the same. However, the result of the method without post-processing can have a relatively large deviation, as shown in Fig. 8(b). This result indicates that our proposed post-processing algorithm can remove some outliers.

3.3. BMO detection performance

To evaluate our proposed method, we compared our detection results with those obtained by the TBMR method proposed by Hussain et al. [17], the patch searching method proposed by Wu et al. [19], and the machine-learning graph-based method proposed by Miri et al. [20]. The mean error of the BMO detection was used as the metrics.

Tables Icon

Table 3. Results of BMO detection

Table 3 shows the mean error of the methods mentioned above and our proposed method compared with the ground truth. Our proposed method obtains better results than the three previously proposed methods, and the results of standard deviation indicate that the stability of our proposed method is superior that TBMR method and the patch searching method, and close to the machine-learning graph-based method. The qualitative evaluation of our proposed method is shown in Fig. 9, and the BMO in the SD-OCT volume is detected and displayed in the 2D projection image. BMO is the true position of the optic disc, and the results are highly consistent with the ground truth.

 figure: Fig. 9

Fig. 9 Comparison of our proposed method (yellow) with the ground truth (red) in the 2D projection image.

Download Full Size | PDF

4. Discussion

To improve the precision of BMO detection and decrease the effect of border tissue and vessel shadows, we present a method for automatic BMO detection in SD-OCT volumes. Compared with other methods, our proposed method has several advantages. First, a three-stage detection framework is used. In the coarse detection, the RPE layer segmentation and the registration between the color fundus image and 2D projection image are employed. In the fixed detection, a U-net is utilized, and an area bias loss function is designed to achieve a high accuracy. Then, a post-processing algorithm is proposed to remove the outliers caused by the border tissue and further improve the performance. Second, instead of searching the entire SD-OCT B-scan, BMO is detected in the ROI determined by the stage of coarse detection. This strategy significantly improves the efficiency and accuracy of our proposed method. Third, the detection of a separate point is transformed to a region segmentation problem by marking the circular region centered on the BMO so that the problem of the imbalanced class is solved. Additional neighborhood information is sent to the model to guide the network learning to discriminate the BMO and the end of the border tissue.

Experimental results prove that the proposed method performs better than existing approaches and can thus be a clinical tool for BMO detection. However, this study also has some limitations. On the one hand, the influence of the vessel shadows and the border tissue caused by peripapillary atrophy cannot be completely eliminated. Thus, a robust compensation algorithm beyond the adaptive compensation [30] is needed to decrease the effects of vessel shadows and peripapillary atrophy. On the other hand, as the ROI is determined by the stage of coarse detection, our proposed method can only deal with the 3D SD-OCT volume because the coarse detection method of single B-scans is not robust enough. With a robust method to detect the ROI of BMO points in a single 2D OCT image, our proposed method can be applied to 2D OCT images by only replacing the coarse detection method.

It is worth noticing that the disc segmentation and registration in our coarse segmentation is not always satisfied enough. In this case, we also use the inflection point of the RPE layer as the coarse BMO points. Although the accuracy will decrease slightly, it will increase the stability of our proposed method. In future works, the ROIs could be extracted by region proposal network based method instead of the coarse detection.

5. Conclusion

In summary, we proposed a method for detecting the BMO points from SD-OCT volumes. The method contains a coarse detection stage, which is composed by an RPE layer segmentation and a multi-modal registration; a fixed detection stage based on the U-net; and a post-processing strategy to further improve the performance by decreasing the effects of the border tissue. Our proposed method enables accurate detection of BMO, leading to precise computation of the parameters based on the BMO points and accurate glaucoma diagnosis.

In future research, we plan to further improve our method in two ways. The network architecture will be modified to suit our detection goal, and a compensation method will be proposed to remove the vessel shadows and increase the contrast of the border tissue and RPE layer.

Funding

National Natural Science Foundation of China (61672542, 81670859); Fundamental Research Funds for the Central Universities of Central South University (2018zzts567).

Disclosures

The authors declare that there are no conflicts of interest related to this article.

References

1. R. N. Weinreb and P. T. Khaw, “Primary open-angle glaucoma,” The Lancet 363, 1711–1720 (2004). [CrossRef]  

2. S. Yousefi, M. H. Goldbaum, M. Balasubramanian, T. Jung, R. N. Weinreb, F. A. Medeiros, L. M. Zangwill, J. M. Liebmann, C. A. Girkin, and C. Bowd, “Glaucoma progression detection using structural retinal nerve fiber layer measurements and functional visual field points,” IEEE Transactions on Biomedical Engineering 61, 1143–1154 (2014). [CrossRef]   [PubMed]  

3. J.-C. Mwanza, R. T. Chang, D. L. Budenz, M. K. Durbin, M. G. Gendy, W. Shi, and W. J. Feuer, “Reproducibility of peripapillary retinal nerve fiber layer thickness and optic nerve head parameters measured with cirrus hd-oct in glaucomatous eyes,” Investigative Ophthalmology & Visual Science 51, 5724 (2010). [CrossRef]  

4. A. Manassakorn, K. Nouri-Mahdavi, and J. Caprioli, “Comparison of retinal nerve fiber layer thickness and optic disk algorithms with optical coherence tomography to detect glaucoma,” American Journal of Ophthalmology 141, 105 –115 (2006). [CrossRef]   [PubMed]  

5. A. Belghith, C. Bowd, F. A. Medeiros, N. Hammel, Z. Yang, R. N. Weinreb, and L. M. Zangwill, “Does the location of bruch’s membrane opening change over time? longitudinal analysis using san diego automated layer segmentation algorithm (salsa),” Investigative Ophthalmology & Visual Science 57, 675 (2016). [CrossRef]  

6. B. C. Chauhan, N. O’Leary, F. A. AlMobarak, A. S. Reis, H. Yang, G. P. Sharpe, D. M. Hutchison, M. T. Nicolela, and C. F. Burgoyne, “Enhanced detection of open-angle glaucoma with an anatomically accurate optical coherence tomography–derived neuroretinal rim parameter,” Ophthalmology 120, 535–543 (2013). [CrossRef]  

7. J. M. Gmeiner, W. A. Schrems, C. Y. Mardin, R. Laemmer, F. E. Kruse, and L. M. Schrems-Hoesl, “Comparison of bruch’s membrane opening minimum rim width and peripapillary retinal nerve fiber layer thickness in early glaucoma assessment,” Investigative Ophthalmology & visual science 57, OCT575 (2016). [CrossRef]  

8. Z.-L. Chen, P. Peng, B.-J. Zou, H.-L. Shen, H. Wei, and R.-C. Zhao, “Automatic anterior lamina cribrosa surface depth measurement based on active contour and energy constraint,” Journal of Computer Science and Technology 32, 1214–1221 (2017). [CrossRef]  

9. Y. Kim, D. Kim, J. Jeoung, D. Kim, and K. Park, “Peripheral lamina cribrosa depth in primary open-angle glaucoma: a swept-source optical coherence tomography study of lamina cribrosa,” Eye 29, 1368 (2015). [CrossRef]   [PubMed]  

10. F. Yin, J. Liu, S. H. Ong, Y. Sun, D. W. Wong, N. M. Tan, C. Cheung, M. Baskaran, T. Aung, and T. Y. Wong, “Model-based optic nerve head segmentation on retinal fundus images,” in Engineering in Medicine and Biology Society, EMBC, 2011 Annual International Conference of the IEEE, (IEEE, 2011), pp. 2626–2629.

11. J. Cheng, J. Liu, Y. Xu, F. Yin, D. W. K. Wong, N.-M. Tan, D. Tao, C.-Y. Cheng, T. Aung, and T. Y. Wong, “Superpixel classification based optic disc and optic cup segmentation for glaucoma screening,” IEEE transactions on Medical Imaging 32, 1019–1032 (2013). [CrossRef]   [PubMed]  

12. Y. Zheng, D. Stambolian, J. O’Brien, and J. C. Gee, “Optic disc and cup segmentation from color fundus photograph using graph cut with priors,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, (Springer, 2013), pp. 75–82.

13. H. Fu, J. Cheng, Y. Xu, D. W. K. Wong, J. Liu, and X. Cao, “Joint optic disc and cup segmentation based on multi-label deep network and polar transformation,” arXiv preprint arXiv:1801.00926 (2018).

14. Z. Hu, M. D. Abramoff, Y. H. Kwon, K. Lee, and M. K. Garvin, “Automated segmentation of neural canal opening and optic cup in 3d spectral optical coherence tomography volumes of the optic nerve head,” Investigative ophthalmology & visual science 51, 5708–5717 (2010). [CrossRef]  

15. M. S. Miri, M. D. Abràmoff, K. Lee, M. Niemeijer, J.-K. Wang, Y. H. Kwon, and M. K. Garvin, “Multimodal segmentation of optic disc and cup from sd-oct and color fundus photographs using a machine-learning graph-based approach,” IEEE Transactions on Medical Imaging 34, 1854–1866 (2015). [CrossRef]   [PubMed]  

16. H. Fu, D. Xu, S. Lin, D. W. K. Wong, and J. Liu, “Automatic optic disc detection in oct slices via low-rank reconstruction,” IEEE Transactions on Biomedical Engineering 62, 1151–1158 (2015). [CrossRef]  

17. A. Belghith, C. Bowd, R. N. Weinreb, and L. M. Zangwill, “A hierarchical framework for estimating neuroretinal rim area using 3d spectral domain optical coherence tomography (sd-oct) optic nerve head (onh) images of healthy and glaucoma eyes,” in Engineering in Medicine and Biology Society (EMBC), 2014 36th Annual International Conference of the IEEE, (IEEE, 2014), pp. 3869–3872.

18. M. A. Hussain, A. Bhuiyan, and K. Ramamohanarao, “Disc segmentation and bmo-mrw measurement from sd-oct image using graph search and tracing of three bench mark reference layers of retina,” in Image Processing (ICIP), 2015 IEEE International Conference on, (IEEE, 2015), pp. 4087–4091. [CrossRef]  

19. M. Wu, T. Leng, L. de Sisternes, D. L. Rubin, and Q. Chen, “Automated segmentation of optic disc in sd-oct images and cup-to-disc ratios quantification by patch searching-based neural canal opening detection,” Optics Express 23, 31216–31229 (2015). [CrossRef]   [PubMed]  

20. M. S. Miri, M. D. Abràmoff, Y. H. Kwon, M. Sonka, and M. K. Garvin, “A machine-learning graph-based approach for 3d segmentation of bruch’s membrane opening from glaucomatous sd-oct volumes,” Medical image analysis 39, 206–217 (2017). [CrossRef]  

21. L. Fang, D. Cunefare, C. Wang, R. H. Guymer, S. Li, and S. Farsiu, “Automatic segmentation of nine retinal layer boundaries in oct images of non-exudative amd patients using deep learning and graph search,” Biomedical Optics Express 8, 2732–2744 (2017). [CrossRef]   [PubMed]  

22. F. G. Venhuizen, B. van Ginneken, B. Liefers, F. van Asten, V. Schreur, S. Fauser, C. Hoyng, T. Theelen, and C. I. Sánchez, “Deep learning approach for the detection and quantification of intraretinal cystoid fluid in multivendor optical coherence tomography,” Biomedical Optics Express 9, 1545–1569 (2018). [CrossRef]   [PubMed]  

23. K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, “Image denoising by sparse 3-d transform-domain collaborative filtering,” IEEE Transactions on image processing 16, 2080–2095 (2007). [CrossRef]   [PubMed]  

24. S. Lu, “Accurate and efficient optic disc detection and segmentation by a circular transformation,” IEEE Transactions on medical imaging 30, 2126–2133 (2011). [CrossRef]   [PubMed]  

25. G. Wang, Z. Wang, Y. Chen, and W. Zhao, “Robust point matching method for multimodal retinal image registration,” Biomedical Signal Processing and Control 19, 68–76 (2015). [CrossRef]  

26. A. Chakravarty and J. Sivaswamy, “A supervised joint multi-layer segmentation framework for retinal optical coherence tomography images using conditional random field,” Computer methods and programs in biomedicine 165, 235–250 (2018). [CrossRef]   [PubMed]  

27. O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, (Springer, 2015), pp. 234–241.

28. F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 3D Vision (3DV), 2016 Fourth International Conference on, (IEEE, 2016), pp. 565–571. [CrossRef]  

29. Y. H. Hwang, J. J. Jung, Y. M. Park, Y. Y. Kim, S. Woo, and J. H. Lee, “Effect of myopia and age on optic disc margin anatomy within the parapapillary atrophy area,” Japanese Journal of Ophthalmology 57, 463–470 (2013). [CrossRef]   [PubMed]  

30. J. M. Mari, N. G. Strouthidis, S. C. Park, and M. J. Girard, “Enhancement of lamina cribrosa visibility in optical coherence tomography images using adaptive compensation,” Investigative ophthalmology & visual science 54, 2238–2247 (2013). [CrossRef]  

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (9)

Fig. 1
Fig. 1 Flowchart of the proposed method.
Fig. 2
Fig. 2 Coarse detection. (a) Original SD-OCT images and 2D projection images. (b) Registration between the projection and fundus images. (c) Diagram of the coarse detection, intersection (yellow) (yellow) of the RPE layer (red), and projection line of the disc (green) are the coarse detection results.
Fig. 3
Fig. 3 Data process.(a) ROI determination. (b) Label transformation for different patches.
Fig. 4
Fig. 4 U-net architecture in our proposed method.
Fig. 5
Fig. 5 Illustration of the border tissue and the post-processing steps. (a) The sketch map of ambiguous border tissue. The real BMO is confused by the border tissue in this case. (b) The sketch map of post-processing. There are four patches (the box in the ROI indicate the upper left patch) extracted in the ROI and send to the trained U-net, the mask of patches with maximum Mi is selected as the result and its center point is the final result of BMO.
Fig. 6
Fig. 6 Results of various radius parameters.
Fig. 7
Fig. 7 Result of various loss function components. (a) Results of group 4 (green), group 2 (yellow), and the ground truth (red). (b) Results of group 4 (green), group 3 (pink), and the ground truth (red). Best viewed in color.
Fig. 8
Fig. 8 Evaluation of the effect of post-processing. The dots in red, green, and yellow represent the ground truth, the result with post-processing, and the result without post-processing, respectively. (a) Result of the method without post-processing is satisfactory. (b) Result of the method without post-processing has a large deviation. Best viewed in color.
Fig. 9
Fig. 9 Comparison of our proposed method (yellow) with the ground truth (red) in the 2D projection image.

Tables (3)

Tables Icon

Table 1 Evaluation of the effect of loss function components

Tables Icon

Table 2 Evaluation of the effect of post-processing

Tables Icon

Table 3 Results of BMO detection

Equations (5)

Equations on this page are rendered with MathJax. Learn more.

D L = 1 2 x Ω p x g x x Ω p x 2 x Ω g x 2
A L = 1 m i n ( x Ω p x , x Ω g x ) m a x ( x Ω p x , x Ω g x )
L o s s = D L + A L + M S E   = 2 2 x Ω p x g x x Ω p x 2 x Ω g x 2 m i n ( x Ω p x , x Ω g x ) m a x ( x Ω p x , x Ω g x ) + 1 n x Ω ( p x g x ) 2
M i = j = 1 , i j m D i c e ( S i , S j ) k
C = C e n t e r ( S d ) , d = arg max  i M i , i = 1 , 2 m
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.