This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: 1 School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
2School of Biomedical Engineering & Imaging Sciences, King’s College London, London, UK
3Department of Radiology, University Hospitals KU Leuven, Leuven, Belgium
4 Department of Obstetrics and Gynaecology, University Hospitals KU Leuven, Leuven, Belgium
5Institute for Women’s Health, University College London, London, UK
6 SenseTime Research, Shanghai, China
11email: [email protected]

Uncertainty-Guided Efficient Interactive Refinement of Fetal Brain Segmentation from Stacks of MRI Slices

Guotai Wang 1122    Michael Aertsen 33    Jan Deprest 44 5 5    Sébastien Ourselin 22    Tom Vercauteren 22    Shaoting Zhang 11 6 6
Abstract

Segmentation of the fetal brain from stacks of motion-corrupted fetal MRI slices is important for motion correction and high-resolution volume reconstruction. Although Convolutional Neural Networks (CNNs) have been widely used for automatic segmentation of the fetal brain, their results may still benefit from interactive refinement for challenging slices. To improve the efficiency of interactive refinement process, we propose an Uncertainty-Guided Interactive Refinement (UGIR) framework. We first propose a grouped convolution-based CNN to obtain multiple automatic segmentation predictions with uncertainty estimation in a single forward pass, then guide the user to provide interactions only in a subset of slices with the highest uncertainty. A novel interactive level set method is also proposed to obtain a refined result given the initial segmentation and user interactions. Experimental results show that: (1) our proposed CNN obtains uncertainty estimation in real time which correlates well with mis-segmentations, (2) the proposed interactive level set is effective and efficient for refinement, (3) UGIR obtains accurate refinement results with around 30% improvement of efficiency by using uncertainty to guide user interactions. Our code is available online111https://github.com/HiLab-git/UGIR.

Keywords:
U

ncertainty \cdot Interactive Segmentation\cdot Fetal brain

1 Introduction

Due to the good soft tissue contrast, fetal Magnetic Resonance Imaging (MRI) is an important tool for diagnosis of abnormalities of the fetal brain during pregnancy [6]. However, MRI is susceptible to motion of fetuses during scanning. To mitigate this problem, fast imaging techniques are often used to obtain stacks of 2D slices that have good in-plane image quality but suffer from large inter-slice motion and low 3D resolution. Segmentation of the fetal brain from stacks of fetal MRI slices plays a critical role for correcting the inter-slice motion and reconstructing a high-resolution 3D volume for fetal brain studies [2, 6, 12].

Despite the fact that deep learning with Convolutional Neural Networks (CNNs) has obtained state-of-the-art performance for automatic fetal brain segmentation from fetal MRI [2, 12], it is still difficult for these automatic segmentation methods to obtain accurate results when dealing with images with motion artifacts, abnormal appearances due to pathologies and some challenging local regions [12]. To address this problem, an efficient interactive method to refine the automatic segmentation result is highly desirable in practice, which makes the segmentation more accurate and robust to be clinically useful [15, 16].

In the literature, some recent works [16, 18] use a second CNN that takes the initial automatic segmentation result and additional user interactions as input to obtain a refined result. In [15], image-specific fine-tuning and Graph Cut [1] were used for interactive refinement. Though these methods achieved higher accuracy and efficiency than traditional interactive segmentation methods [17], when used to segment a volumetric data, they require the user to carefully check the initial segmentation in 2D views slice-by-slice and manually identify mis-segmented regions to give interactions for refinement. For stacks of fetal MRI slices, automatic CNNs could obtain accurate initial segmentation for most slices [2, 12], manually identifying mis-segmented regions may be unnecessary for accurately segmented slices while sometimes difficult for challenging slices. Therefore, the efficiency of such methods is limited.

To improve the efficiency for user interactions, leveraging the uncertainty information of the initial segmentation has been shown to be a promising method [13] as it can automatically identify potential mis-segmentations and guide the user to give interactions only in some uncertain regions. Despite the availability of several reliable uncertainty estimation methods for CNN-based segmentation, such as Monte Carlo (MC) dropout [3], model ensemble [8, 5] and test-time augmentation [14], they require multiple forward passes at inference time and cannot provide real-time uncertainty estimation for guiding user interactions. Alternatively, a Bayesian Network has been proposed for fast uncertainty estimation with a single forward pass [4]. However, their utility for guiding interactive refinement has not been investigated.

In this paper, we propose a novel uncertainty-guided framework for interactive refinement of automatic segmentation obtained by CNNs and apply it to fetal brain segmentation from fetal MRI. The contribution is three-fold. First, to obtain real-time uncertainty estimation, we propose a novel method using Grouped Convolution (GC)-based CNNs. It obtains multiple predictions simultaneously with a single model and gives uncertainty estimation in a single forward pass, which is more suitable in the scenario of interactive refinement. Second, we propose to guide the user to give interactions more efficiently during refinement according to the uncertainty information. Thirdly, we propose a novel interactive level set method that incorporates the initial segmentation and user interactions in a uniformed framework to obtain accurate refined results efficiently. The superiority of our framework over existing methods was validated in the task of fetal brain segmentation from stacks of motion-corrupted fetal MRI slices.

2 Methods

Refer to caption
Figure 1: The proposed Uncertainty-Guided Interactive Refinement (UGIR) framework for fetal brain segmentation. For an input stack with MM slices, the user only needs to give interactions in a subset of M(M<M)M^{\prime}(M^{\prime}<M) slices with the highest uncertainty.

Our proposed Uncertainty-Guided Interactive Refinement (UGIR) framework for efficient interactive fetal brain segmentation is shown in Fig. 1. First, a novel CNN based on convolution in Multiple Groups (MG-Net) simultaneously obtains the initial segmentation and uncertainty estimation of a stack of MRI slices in real time. Let SmS_{m} (mm = 1, 2, …) denote the slice with the mm-th highest uncertainty. Our framework automatically and iteratively fetches slice SmS_{m} as suggestion for user interactions. After the user gives interactions in SmS_{m}, a novel interactive level set method obtains the refined result of SmS_{m}. The iterative refinement is finished when no further slice is suggested by the framework.

2.0.1 Simultaneous Initial Segmentation and Uncertainty Estimation.

Refer to caption
Figure 2: The proposed MG-Net that obtains NN (e.g., 4) predictions corresponding to NN groups of features. The numbers over each feature map such as N×CN\times C are group number ×\times channel number of each group. GC(N) denotes convolution with NN groups.

We propose MG-Net to obtain simultaneous initial segmentation and uncertainty estimation in real time with a single network and a single forward pass for inference, which is more efficient for interactive segmentation than typical uncertainty estimation methods including MC dropout [3] and model ensemble [5, 8]. As shown in Fig. 2, we modify U-Net [11] by using Grouped Convolution [7].

An NN-grouped convolution layer splits the input feature map along the channel dimension into NN groups each with CiC_{i} channels. For each group, it uses a convolution kernel of shape Co×Ci×h×wC_{o}\times C_{i}\times h\times w respectively, where h×wh\times w is the spatial size of the kernel. Therefore, NN independent feature maps are obtained each with CoC_{o} channels. They are concatenated into a single one with N×CoN\times~{}C_{o} channels. Note that there is no correlation between different groups in the output feature map of an NN-grouped convolution layer. Similarly, we implement up-sampling layers by transposed NN-grouped convolutions, and extend standard channel concatenation to group-wise concatenation to keep the NN groups independent of each other. Let F1F_{1} and F2F_{2} represent two feature maps each with NN groups. We first concatenate the nn-th group of F1F_{1} with the nn-th group of F2F_{2}, and denote the result as F^n\hat{F}_{n}. Then F^1\hat{F}_{1}, F^2\hat{F}_{2}, …, and F^N\hat{F}_{N} are concatenated as the group-wise concatenation of F1F_{1} and F2F_{2}. The group-wise concatenation is used for skip connection between feature maps in the encoder and their counterparts in the decoder, as shown in Fig. 2. At the last layer of the decoder, we apply softmax to each feature group respectively (i.e., group-wise softmax) to obtain NN probability predictions. Therefore, MG-Net can be seen as an ensemble of NN parallel sub-networks, and they are randomly initialized and trained with dropout to obtain diversity. At the lowest resolution level of MG-Net, we set group number to one to allow communication of these NN sub-networks for better performance.

For training, we apply the same segmentation loss function (i.e., Dice loss [10]) to the NN predictions respectively and average their loss values for back-propagation. At inference time, we take the pixel-wise mean value and variance of the NN probability predictions as the final segmentation probability map and pixel-level uncertainty estimation, respectively. The mean foreground probability map is thresholded by 0.5 to obtain a binary segmentation mask.

2.0.2 Uncertainty-Guided User Interactions.

For a stack of MM slices, the CNN is able to obtain accurate segmentation for most slices and only a few slices may require manual refinement [12]. To avoid unnecessary and time-consuming manual search for mis-segmentations of each slice, we ask the user to check and refine only MM^{\prime} (M<MM^{\prime}<M) slices with the highest slice-level uncertainty. For a slice SS, its binary segmentation result YY and pixel-level uncertainty map UU, a naive slice-level uncertainty can be defined as ν=𝐱U𝐱\nu^{*}=\sum_{\mathbf{x}}U_{\mathbf{x}} where U𝐱U_{\mathbf{x}} is the uncertainty of pixel 𝐱\mathbf{x}. However, this may lead a slice with a small target to be neglected as it often has a low value of ν\nu^{*} due to a small uncertain region. To address this problem, we alternatively define the slice-level uncertainty as ν=(𝐱U𝐱)/(𝐱Y𝐱+ζ)\nu=(\sum_{\mathbf{x}}U_{\mathbf{x}})/(\sum_{\mathbf{x}}Y_{\mathbf{x}}+\zeta), which is normalized by the segmented region size. ζ\zeta is a small number for numerical stability. We iteratively fetch the slice SmS_{m} with the mm-th highest ν\nu to ask for interactive refinement, where mm = 1, 2, …, and the iteration terminates when m>Mm>M^{\prime}, as shown in Fig. 1. MM^{\prime} is a predefined number, such as 60% of MM according to the performance of the CNN on the validation set. We also use an early termination strategy when refinement is not needed for three consecutive fetched slices.

2.0.3 Interaction-based Level Set for Fast Refinement.

For interactive refinement, we use a Distance Regularized Level Set Evolution (DRLSE) [9] due to its efficiency and extend it with an interaction-constraint term, which is named as I-DRLSE. Let ϕ\phi denote the level set function that is initialized as the signed distance transform of the initial segmentation result. We define an energy function as E(ϕ)=αEr+βEu+λEl+μEdE(\phi)=\alpha E_{r}+\beta E_{u}+\lambda E_{l}+\mu E_{d}, where α\alpha, β\beta, λ\lambda, μ\mu are weighting parameters. ErE_{r}, EuE_{u}, ElE_{l} and EdE_{d} are the region, user-interaction, length and distance regularization terms, respectively. El=Ωδϵ(ϕ(𝐱))|ϕ(𝐱)|𝑑𝐱E_{l}=\int_{\Omega}\delta_{\epsilon}(\phi(\mathbf{x}))|\nabla\phi(\mathbf{x})|d\mathbf{x} and Ed=Ωp(|ϕ(𝐱)|)𝑑𝐱E_{d}=\int_{\Omega}p(|\nabla\phi(\mathbf{x})|)d\mathbf{x}, where δϵ\delta_{\epsilon} and p()p() are the smoothed Dirac delta function and double-well potential function as in [9] respectively. As the target region in the fetal MRI image has an inhomogeneous appearance that brings challenges to standard intensity-based level set methods, we define the region term based on the foreground probability map PP obtained by MG-Net instead of the original image:

Er=Ω(|Pc1|2Hϵ(ϕ(𝐱))+|Pc2|2(Hϵ(ϕ(𝐱))))𝑑𝐱\displaystyle E_{r}=\int_{\Omega}\Big{(}|P-c_{1}|^{2}H_{\epsilon}(\phi(\mathbf{x}))+|P-c_{2}|^{2}(H_{\epsilon}(-\phi(\mathbf{x})))\Big{)}d\mathbf{x} (1)

where HϵH_{\epsilon} is the smoothed Heaviside function as in [9]. c1c_{1} and c2c_{2} are the average foreground probability inside and outside the current level set contour, respectively. Our proposed user-interaction term is:

Eu=Ω(Hϵ(ϕ(𝐱))log(η(𝐱))+Hϵ(ϕ(𝐱))log(1η(𝐱)))𝑑𝐱\displaystyle E_{u}=-\int_{\Omega}\Big{(}H_{\epsilon}(\phi(\mathbf{x}))\text{log}(\eta(\mathbf{x}))+H_{\epsilon}(-\phi(\mathbf{x}))\text{log}(1-\eta(\mathbf{x}))\Big{)}d\mathbf{x} (2)

where η(𝐱)\eta(\mathbf{x}) is user-interaction-derived likelihood of pixel 𝐱\mathbf{x} being the foreground. Let \mathcal{F} and \mathcal{B} represent the set of pixels specified as the foreground and background by the user interactions, respectively. Inspired by [16], we use g𝐱g_{\mathbf{x}}^{\mathcal{F}} to represent the geodesic distance between 𝐱\mathbf{x} and \mathcal{F}, and define η(𝐱)=eG𝐱/(eG𝐱+eG𝐱)\eta(\mathbf{x})=e^{-G_{\mathbf{x}}^{\mathcal{F}}}/(e^{-G_{\mathbf{x}}^{\mathcal{F}}}+e^{-G_{\mathbf{x}}^{\mathcal{B}}}), where G𝐱=min(g𝐱,D)G_{\mathbf{x}}^{\mathcal{F}}=\min(g_{\mathbf{x}}^{\mathcal{F}},D) and G𝐱=min(g𝐱,D)G_{\mathbf{x}}^{\mathcal{B}}=\min(g_{\mathbf{x}}^{\mathcal{B}},D) and DD is a threshold value to ensure that only a local region is affected by the interactions. We set g𝐱g_{\mathbf{x}}^{\mathcal{F}} or g𝐱g_{\mathbf{x}}^{\mathcal{B}} as DD when \mathcal{F} or \mathcal{B} is empty. EuE_{u} is infinite if the segmentation results conflict with the user interactions.

3 Experiments and Results

3.0.1 Data and Implementation.

MRI scans of 35 fetuses in the second trimester were collected by Single Shot Fast Spin Echo (SSFSE) with pixel size 0.74 mm-1.58 mm and inter-slice spacing 3 mm-4 mm. Each fetus had three scans in axial, sagittal and coronal views respectively, leading to 105 stacks in total. We randomly selected 72 stacks from 24 patients, 9 stacks from 3 patients and 24 stacks from 8 patients for training, validation and testing, respectively. Manual segmentations of these images were used as the ground truth.

We implemented the CNNs in Pytorch on a Ubuntu desktop with an NVIDIA GTX 1080 Ti GPU and developed a PyQt GUI for user interactions. I-DRLSE was implemented in Python and it ran on the CPU. To train MG-Net, we used Dice loss [10] and Adam optimizer with weight decay 10510^{-5}, mini-batch of 24 slices and learning rate 10410^{-4}. The training was ended when performance on the validation set stopped to increase for 5k iterations. The group number NN in MG-Net was 4 and the channel number parameter CC in Fig. 2 was 16. For interactive refinement, MM^{\prime} was set to 0.6M0.6M as refinement was not needed for more than half of the slices in the validation set. For I-DRLSE, the maximal evolution step number was 200 and α\alpha=0.1, β\beta=0.5, λ\lambda=0.3, μ\mu=0.005, DD=4.0 based on grid search on the validation set. The segmentation accuracy was measured by Dice similarity and Average Symmetric Surface Distance (ASSD) between segmentation results and the ground truth.

3.0.2 Initial Segmentation and Uncertainty Estimation.

Table 1: Quantitative evaluation of different uncertainty estimation methods for fetal brain segmentation before refinement. The values were measured at stack-level.
Method Efficiency Uncertainty Quality Segmentation Quality
Param(M) Runtime(s) UEO(%) RVE(%) Dice(%) ASSD(mm)
U-Net [11] 11.51 0.31±\pm0.09 91.54±\pm5.87 4.31±\pm2.45
MC Dropout [3] 11.51 3.12±\pm0.83 36.17±\pm5.31 42.64±\pm36.94 92.04±\pm9.27 4.16±\pm2.75
Ensemble [8] 11.51 1.56±\pm0.42 37.08±\pm7.75 40.74±\pm37.29 92.05±\pm5.72 4.07±\pm2.37
Bayesian Net [4] 11.51 0.31±\pm0.09 36.64±\pm5.10 26.54±\pm18.29 90.24±\pm7.39 4.87±\pm2.82
MG-Net 4.86 0.29±\pm0.08 40.47±\pm3.58 21.90±\pm17.62 91.82±\pm4.78 4.07±\pm2.13
Refer to caption
Figure 3: Comparison of different methods for initial segmentation (1st row) and uncertainty estimation (2nd row) that is normalized to [0, 1]. Red arrows highlight the consistency between mis-segmentation (eye ball in (a) and lung in (b)) and uncertainty obtained by MG-Net.

Our MG-Net was compared with MC dropout [3] with 10 folds, ensemble [8] of 5 models, and a Bayesian network [4] for uncertainty estimation. We implemented all these methods using U-Net [11] as the backbone. To measure ability of the uncertainty to indicate mis-segmentation, we used Uncertainty-Error Overlap (UEO, i.e., Dice as in [5]) and Relative Volume Error (RVE) between thresholded uncertain region and mis-segmented region. The optimal threshold value for each method was determined based on the validation set. Quantitative evaluation results are shown in Table 1. Compared with ensemble of U-Net, our MG-Net obtains comparable segmentation accuracy and higher uncertainty estimation quality. MG-Net takes 0.29s in average for simultaneous automatic segmentation and uncertainty estimation of a stack, which is far more efficient than MC dropout and model ensemble and more suitable for interactive segmentation. The Bayesian network method also obtains uncertainty estimation in a fast speed, but with a reduced segmentation accuracy. Visual comparison in Fig. 3 shows that MG-Net obtains better consistency between mis-segmentation and uncertain regions than the other uncertainty estimation methods.

3.0.3 Refinement using Interactive Level Set.

Our I-DRLSE was firstly validated with slice-level refinement, and compared with: 1) CPU-based Graph Cut222Code from the maxflow-v3.01 library: https://vision.cs.uwaterloo.ca/code/ as implemented in [15], and 2) training a refinement CNN (a.k.a., R-Net) [16] that takes the initial segmentation and user interactions as input. We implemented the R-Net using U-Net [11] as the backbone, and trained it with simulated interactions on initial segmentation obtained by MG-Net following [16]. As not all the slices in a stack require refinement, we randomly selected 100 obviously mis-segmented slices from the test set, and used the same set of user interactions on the initial segmentation for comparison. Fig. 4(a) shows two cases where over- and under-segmentation exist in the initial segmentation respectively, which demonstrates that I-DRLSE obtains higher refinement accuracy than Graph Cut and R-Net with the same initial segmentation and interactions. Fig. 4(b) shows a quantitative comparison of these refinement methods. It can be observed that I-DRLSE leads to higher Dice scores and lower ASSD values than Graph Cut and R-Net for refinement. The average slice-level machine time for CPU-based I-DRLSE was 0.45s, which is slower than that of Graph Cut (0.07s) and R-Net (0.12s), but still acceptable for fast response of user interactions. The efficiency of I-DRLSE could be further improved by a GPU-optimized implementation in the future.

Refer to caption
Figure 4: Slice-level qualitative and quantitative comparisons of different methods for refinement based on the same set of initial segmentation and user interactions.

3.0.4 Comparison of Different Interactive Frameworks.

For stack-level segmentation, our UGIR was compared with two variants: 1) UGIR(-U) denoting that the user manually searches mis-segmentation slice-by-slice to provide interactions (i.e., not guided by uncertainty) for refinement using I-DRLSE; 2) UGIR(*) that denotes using the naive slice-level uncertainty ν\nu^{*} to guide user interactions. They were also compared with two existing interactive segmentation methods: using Graph Cut [1] for interactive segmentation from scratch, and DeepIGeoS [16] that uses two CNNs for initial segmentation and interactive refinement, respectively. We re-implemented DeepIGeoS following the training method in [16]. One user employed these methods to segment the fetal brain from the testing fetal MRI stacks respectively, where the interactions in a slice could be given multiple times until the result was accepted. Quantitative evaluation results are shown in Table 2. Compared with DeepIGeoS [16], our UGIR obtained similar final accuracy (pp-value >> 0.05 based on a paired t-test), but reduced the runtime from 75.38s to 48.46s. By using uncertainty-guided user interactions, UGIR improved the efficiency by near 30% from UGIR(-U). UGIR(*) and UGIR took almost the same runtime, but UGIR achieved higher accuracy.

Table 2: Quantitative comparison of different interactive methods for stack-level segmentation of the fetal brain.
Graph Cut [1] DeepIGeoS [16] UGIR(-U) UGIR(*) UGIR
Dice (%) 93.17±\pm3.15 95.03±\pm3.07 95.00±\pm3.09 94.65±\pm3.28 94.83±\pm3.22
ASSD (mm) 2.84±\pm1.18 2.72±\pm1.74 2.73±\pm1.12 2.75±\pm1.17 2.70±\pm1.15
Runtime (s) 214.54±\pm58.73 75.38±\pm42.67 68.04±\pm26.00 48.11±\pm20.87 48.46±\pm19.40

4 Conclusion

In this work, we propose a novel interactive segmentation framework using uncertainty to efficiently guide user interactions for refining results obtained by automatic CNNs. We introduce MG-Net based on grouped convolution to obtain multiple segmentation predictions simultaneously with real-time uncertainty estimation, which is used to suggest mis-segmented slices for user interactions, avoiding unnecessary manual check of well-segmented slices and leading to improved efficiency. A novel interactive level set I-DRLSE is also proposed to obtain refined results with spatial regularization. Experiments with fetal brain segmentation from stacks of motion-corrupted fetal MRI slices show that the proposed interactive framework achieved high accuracy with fast runtime, and the uncertainty information helped to improve the refinement efficiency by around 30%.

4.0.1 Acknowledgements.

This work was supported by the National Natural Science Foundation of China funding [81771921, 61901084], the Wellcome Trust [WT101957, 203148/Z/16/Z], and the Engineering and Physical Sciences Research Council (EPSRC) [NS/A000027/1, NS/A000049/1]. TV is supported by a Medtronic / Royal Academy of Engineering Research Chair [RCSRF1819/7/34].

References

  • [1] Boykov, Y.Y., Jolly, M.P.: Interactive graph cuts for optimal boundary & region segmentation of objects in N-D images. In: ICCV. pp. 105–112 (2001)
  • [2] Ebner, M., Wang, G., Li, W., Aertsen, M., Patel, P.A., Aughwane, R., Melbourne, A., Doel, T., Dymarkowski, S., De Coppi, P., David, A.L., Deprest, J., Ourselin, S., Vercauteren, T.: An automated framework for localization, segmentation and super-resolution reconstruction of fetal brain MRI. Neuroimage 206, 116324 (2020)
  • [3] Gal, Y., Ghahramani, Z.: Dropout as a Bayesian approximation: representing model uncertainty in deep learning. In: ICML. pp. 1050–1059 (2016)
  • [4] Jena, R., Awate, S.P.: A bayesian neural net to segment images with uncertainty estimates and good calibration. In: IPMI. pp. 3–15 (2019)
  • [5] Jungo, A., Reyes, M.: Assessing reliability and challenges of uncertainty estimations for medical image segmentation. In: MICCAI. vol. 1, pp. 48–56 (2019)
  • [6] Keraudren, K., Kuklisova-Murgasova, M., Kyriakopoulou, V., Malamateniou, C., Rutherford, M.A., Kainz, B., Hajnal, J.V., Rueckert, D.: Automated fetal brain segmentation from 2D MRI slices for motion correction. Neuroimage 101, 633–643 (2014)
  • [7] Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: NeurIPS. pp. 1097–1105 (2012)
  • [8] Lakshminarayanan, B., Pritzel, A., Blundell, C.: Simple and scalable predictive uncertainty estimation using deep ensembles. In: NeurIPS. pp. 6405–6416 (2017)
  • [9] Li, C., Xu, C., Gui, C., Fox, M.D.: Distance regularized level set evolution and its application to image segmentation. IEEE Trans. Image Process. 19(12), 3243–3254 (2010)
  • [10] Milletari, F., Navab, N., Ahmadi, S.A.: V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In: IC3DV. pp. 565–571 (2016)
  • [11] Ronneberger, O., Fischer, P., Brox, T.: U-Net: Convolutional networks for biomedical image segmentation. In: MICCAI. pp. 234–241 (2015)
  • [12] Salehi, S.S.M., Hashemi, S.R., Velasco-Annis, C., Ouaalam, A., Estroff, J.A., Erdogmus, D., Warfield, S.K., Gholipour, A.: Real-time automatic fetal brain extraction in fetal MRI by deep learning. In: ISBI. pp. 720–724 (2018)
  • [13] Top, A., Hamarneh, G., Abugharbieh, R.: Active learning for interactive 3D image segmentation. In: MICCAI. pp. 603–610 (2011)
  • [14] Wang, G., Li, W., Aertsen, M., Deprest, J., Ourselin, S., Vercauteren, T.: Aleatoric uncertainty estimation with test-time augmentation for medical image segmentation with convolutional neural networks. Neurocomputing 338, 34–45 (2019)
  • [15] Wang, G., Li, W., Zuluaga, M.A., Pratt, R., Patel, P.A., Aertsen, M., Doel, T., David, A.L., Deprest, J., Ourselin, S., Vercauteren, T.: Interactive medical image segmentation using deep learning with image-specific fine-tuning. IEEE Trans. Med. Imaging 37(7), 1562–1573 (2018)
  • [16] Wang, G., Zuluaga, M.A., Li, W., Pratt, R., Patel, P.A., Aertsen, M., Doel, T., David, A.L., Deprest, J., Ourselin, S., Vercauteren, T.: DeepIGeoS: A deep interactive geodesic framework for medical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 41(7), 1559–1572 (2019)
  • [17] Zhao, F., Xie, X.: An overview of interactive medical image segmentation. Ann. BMVA 2013(7), 1–22 (2013)
  • [18] Zhou, B., Chen, L., Wang, Z.: Interactive deep editing framework for medical image segmentation. In: MICCAI. pp. 329–337 (2019)