This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Uncertainty-Aware Deep Co-training for Semi-supervised Medical Image Segmentation

Xu Zheng Chong Fu11footnotemark: 122footnotemark: 233footnotemark: 3 fuchong Haoyu Xie Jialei Chen Xingwei Wang Chiu-Wing Sham School of Computer Science and Engineering, Northeastern University, Shenyang 110819, China Engineering Research Center of Security Technology of Complex Network System, Ministry of Education, China Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang 110819, China School of Computer Science, The University of Auckland, New Zealand
Abstract

Semi-supervised learning has made significant strides in the medical domain since it alleviates the heavy burden of collecting abundant pixel-wise annotated data for semantic segmentation tasks. Existing semi-supervised approaches enhance the ability to extract features from unlabeled data with prior knowledge obtained from limited labeled data. However, due to the scarcity of labeled data, the features extracted by the models are limited in supervised learning, and the quality of predictions for unlabeled data also cannot be guaranteed. Both will impede consistency training. To this end, we proposed a novel uncertainty-aware scheme to make models learn regions purposefully. Specifically, we employ Monte Carlo Sampling as an estimation method to attain an uncertainty map, which can serve as a weight for losses to force the models to focus on the valuable region according to the characteristics of supervised learning and unsupervised learning. Simultaneously, in the backward process, we joint unsupervised and supervised losses to accelerate the convergence of the network via enhancing the gradient flow between different tasks. Quantitatively, we conduct extensive experiments on three challenging medical datasets. Experimental results show desirable improvements to state-of-the-art counterparts.

keywords:
Semi-supervised Learning , Co-training , Uncertainty , Medical Image Segmentation
\useunder

\ul

Refer to caption
Figure 1: Overview of the uncertainty-aware co-training algorithm. On the basis of training two deep CNN models simultaneously with two different sets of the labeled datasets and the same unlabeled dataset, we bring in Monte Carlo Sampling to acquire uncertainty, which is acting on the loss function in both the supervised learning stage and the unsupervised learning stage.

1 Introduction

Semantic segmentation is a fundamental computer vision task that is particularly important for medical image analysis [1]. It serves as a preprocessing step for the treatment of different medical conditions. In recent years, convolutional neural network (CNN) based methods have achieved tremendous improvement [2][3]. With abundant pixel-wise labeled data, CNN can easily achieve outstanding performance relying on great nonlinear fitting capability [4]. However, pixel-wise annotated images are expensive to obtain especially in the medical domain. It is time-consuming and extremely relies on experienced experts. Fortunately, semi-supervised methods leverage not only labeled data but also unlabeled data [5], freeing the researchers from labeling work. Therefore, semi-supervised learning has attracted great attention [6][7][8].

The current dominant semi-supervised learning methods in deep learning are self-training [9], consistency regulation [10], pseudo labeling[11], adversarial learning [12] and so on. All the above approaches are based on the assumption that semi-supervised methods rely on intrinsic properties of the dataset distribution rather than individual images. So, parameter optimization can process with a combination of annotated images and unannotated images rather than labeled images only. For example, self-training [10] selects the most reliable prediction of the current model as the label (pseudo label), for the complement of the labeled dataset. Other approaches attempt to exchange internal information among an ensemble of models, aimed at increasing their consistency, such as co-training [13]. These approaches are all constructive to a certain degree. But how to leverage unannotated data more effectively is still one of the most concerning issues in these semi-supervised learning methods.

Most of the existing semi-supervised methods leverage the prediction of unannotated data, but the quality of the prediction is not guaranteed. Optimizing model parameters via unreliable predictions in unsupervised learning is not convinced, even towards wrong results. Here, the major challenge is to assess the reliability of the prediction. To address this issue, we propose an uncertainty-aware semi-supervised semantic segmentation framework, as illustrated in Fig. 1. Uncertainty is an estimation of prediction confidence [14], which is leveraged as prior information to guide network learning by adjusting loss weight dynamically in two separate strategies in our framework. It forces the models to focus on the valuable region according to the characteristics of supervised learning and unsupervised learning. On the one hand, it guides networks to pay more attention to the pixels with high prediction reliability (low uncertainty) in unsupervised learning; on the other hand, it guides networks to concentrate more on ambiguous pixels (high uncertainty) in supervised learning. In addition, owing to its success in semi-supervised learning, deep co-training is utilized in our proposed framework. The general principle of co-training is to independently train models with the labeled data and enforce their predictions to agree on the labeling of the unlabeled data. We separate the labeled dataset into complementary subsets and use adversarial examples [15] to enforce the diversity. Moreover, to promote the predictions consistency among different models, we add ensemble agreement loss to guide models training. We joint unsupervised and supervised losses to accelerate the convergence of the network via enhancing the gradient flow between different tasks.

In this paper, we explore the uncertainty information to enforce the performance of networks both in supervised learning and unsupervised learning. With the guidance of the estimated uncertainty in both supervised and unsupervised stages, our uncertainty-aware deep co-training method achieves the state-of-the-art performance in the semi-supervised tasks on three public medical datasets. Moreover, a comprehensive set of experiments are conducted which demonstrate the potential of our approach for segmenting different types of medical images. Our experiments analyze the impact of various elements of the method.

2 Related work

2.1 Semi-supervised segmentation

Semi-supervised learning is an indispensable part of deep learning theory [16]. Recently, many works have proposed using semi-supervised approaches in medical image segmentation to segment human organs [17]. An innovative semi-supervised segmentation approach is presented by Mohamed et al. for efficient segmentation of lung CT scans from Convid-19 patients. Yu et al. [18] proposed a teacher-student model to segment 3D left atrium via self-ensembling. Li et al. [10] proposed a semi-supervised segmentation method for skin lesions by enforcing the consistency between the student and teacher model. Fang et al. [19] used a co-training framework to boost each sub-model for kidney tumor segmentation. And they applied adversarial training [20] to their network to constrain the models to output invariant results over different perturbations on input data.

Similar to co-training, Peng et al. [13] presented a method based on an ensemble of deep segmentation networks. They trained different models with corresponding subsets of the annotated data and leveraged non-annotated images to exchange internal information among sub-models. The difference between these semi-supervised methods is how they utilize unlabeled data and the way they relate to supervised algorithms [5]. Most of the approaches utilize predictions of unlabeled data for consistency training [21]. However, the model’s prediction of unlabeled data used to exchange information is unstable. We need to add components to enable us to learn the specific confidence situation of each pixel of the prediction [22]. So we introduce uncertainty estimation [23] in our approach.

2.2 Uncertainty Estimation

Uncertainty estimation plays a pivotal role in reducing the impact of stochastic during both optimization and decision-making processes [24]. Knowing the confidence (uncertainty) with which we can trust the neural networks’ predictions is essential for decision making [25]. A mathematically grounded framework is offered by Bayesian probability theory to learn about the uncertainty of a data-driven model, called Bayesian Neural Network(BNN). BNN attempted to learn a distribution over each of the network’s weight parameters [14]. However, Bayesian inference is intractable on computation in practice.

Deep ensemble [26] is an alternative to BNN, which observes the variance of many trained models to estimate uncertainty. In addition, several works applied different uncertainty estimation methods to the task of semantic segmentation, including Stochastic Batch Normalization [27], Multiplicative Normalizing Flows [28] and so on. Upadhyay et al. [29] proposed a GAN-based framework that estimates the per-voxel uncertainty in the predictions on magnetic resonance imaging (MRI). Alex [14] studied models under the framework with per-pixel semantic segmentation and proposed new loss functions based on uncertainty, interpreted as learned attenuation. Wu et al. [21] claimed that deep models require extra components to obtain performance gains. And Yu et al. [18] introduced uncertainty to ensemble models by using Monte-Carlo Sampling. It has been proved that the Monte Carlo Sampling has much superiority over conventional methods in the estimation of uncertainty, especially of complex measurement systems’ output [25].

3 Methods

3.1 Problem formulation

Since manual annotation is often time-consuming and expensive, only a small fraction of images in datasets can have full pixel-wise labels. So our proposed method aims to learn more information from both labeled and abundant unlabeled images by using the uncertainty-aware deep co-training methodology. We formalize the problem of medical image semantic segmentation as follows.

Given labeled dataset Dl={(x1l,y1l),,(xml,yml)}D_{l}={\{(x_{1}^{l},y_{1}^{l}),...,(x_{m}^{l},y_{m}^{l})}\}, which contains mm labeled examples, each example comprised of an input image xilx_{i}^{l} with dimensions H×W{H\times W} and its corresponding pixel-level label yilH×W×Cy_{i}^{l}\in\mathbb{R}^{H\times W\times C}, where CC is the number of classes and H×WH\times W is spatial dimension . In semi-supervised setting, we also have unlabeled dataset Du={x1u,,xnu}D_{u}=\{x_{1}^{u},...,x_{n}^{u}\} including nn unlabeled images, with nmn\gg m. The purpose of semi-supervised segmentation is to train a segmentation model ff of parameter θ\theta with D=DlDuD=D_{l}\cup D_{u}, which map each pixel of an input image to its correct label.

Refer to caption
Figure 2: Illustration of supervised learning. Each model outputs predictions of the image from corresponding labeled subsets, computing supervised loss LsupL_{sup} with labels, illustrated as red lines. Simultaneously, the uncertainty is estimated by Monte Carlo Sampling. Each model approximates its own uncertainty by performing TT stochastic forward passes and multiplies the uncertainty UsupU_{sup} as a weight coefficient to the supervised loss LsupL_{sup}.

3.2 Proposed approach

We train two models f(θ1)f(\theta_{1}), f(θ2)f(\theta_{2}) jointly with labeled dataset DlD_{l} and unlabeled dataset DuD_{u} in a collaborative manner. Motivated by performance improvement from the uncertainty estimation in the Bayesian networks, we employ Monte Carlo Sampling[14] to estimate uncertainty. Specifically, we perform TT stochastic forward passes on two models under random dropout. According to this, we obtain a set of predictions and choose the predictive entropy metric to approximate the uncertainty. The predictive entropy can be summarized as:

μc=1TtPtcandu=cμc𝑙𝑜𝑔μc,\displaystyle\mu_{c}=\frac{1}{T}\sum_{t}P_{t}^{c}\qquad and\qquad u=-\sum_{c}\mu_{c}\mathit{log}\mu_{c}, (1)

where PtcP_{t}^{c} is the probability of the c-th class in the t-th time prediction. And the uncertainty is estimated in pixel level, the uncertainty of the whole volume UU is {u}𝐑H×W×D\left\{u\right\}\in\mathbf{R}^{H\times W\times D}.

Following co-training methods for semi-supervised semantic segmentation, we employ a loss function composed of a weighted sum of three separate terms to train ensemble models.

(θ;D)=sup(θ;Dl)+λcotagr(θ;Du)+λdivdiv(θ;D).\mathcal{L}(\theta;D)=\mathcal{L}_{sup}(\theta;D_{l})+\lambda_{cot}\mathcal{L}_{agr}(\theta;D_{u})+\lambda_{div}\mathcal{L}_{div}(\theta;D). (2)

Uncertainty is used to impact the weight of the first two parts of the loss function. The details are explained in the following subsections.

3.2.1 Supervised learning

As in standard multi-view learning [30] approach, we separate the labeled dataset DlD_{l} to complementary subsets Dl1D_{l_{1}} and Dl2D_{l_{2}}. Training two models individually with corresponding labeled subsets makes each model have a certain fit to the distribution of the data. And also, we can ensure the diversity of the two models.

We employ E-Net[31] as our network backbone. To adapt the E-Net as a Bayesian network to estimate the uncertainty, one dropout layer is added between the encoder and decoder. In the network training and uncertainty estimation, we turn on the dropout layer. Since we don’t need to estimate uncertainty in the testing phase, we turn off the dropout layer.

Through the forward propagation of each model TT times, we can get TT different results as shown in Fig. 2. Therefore, we obtain a set of Softmax probability vectors for each pixel in the input. By using predictive entropy as the metric, we can approximate uncertainty maps. We express the uncertainty map as the model confidence of each pixel prediction. If a pixel has a large fluctuation during the prediction process of TT forward propagation, the uncertainty of it will be large.

With the guidance of the estimated uncertainty Usup1U_{sup_{1}} and Usup2U_{sup_{2}}, we find out the unreliable prediction in images and make models learn more from these pixels from corresponding labels. We design the uncertainty-aware supervised loss LsupL_{sup} as the pixel-level cross-entropy loss of the two models:

sup(θ,Dl)=Usup1Dl1𝐶𝐸(f(xi;θ1),yi)\displaystyle\mathcal{L}_{sup}(\theta,D_{l})=U_{sup_{1}}\sum_{D_{l_{1}}}\mathit{CE}(f(x_{i};\theta_{1}),y_{i}) (3)
+Usup2Dl2𝐶𝐸(f(xi;θ2),yi).\displaystyle+U_{sup_{2}}\sum_{D_{l_{2}}}\mathit{CE}(f(x_{i};\theta_{2}),y_{i}).

3.2.2 Unsupervised learning

Semi-supervised learning means not only exploiting labeled images, unlabeled images DuD_{u} are more important for the learning process. The strategy of unsupervised learning wants the segmentation networks to output similar predictions for the same unlabeled image with perturbations. The deep co-training method minimizes the distance between the class distributions predicted by different models. However, the prediction results obtained through one forward pass are contingent, which will limit models’ ability to fit the unlabeled data distribution.

So, we also introduce the concept of uncertainty in unsupervised learning. Same as the fully supervised phase, we perform TT stochastic forward passes on each model under random dropout as shown in Fig. 3. After leveraging predictive entropy as a metric, we get two uncertainty maps approximated by models. Different from the supervised stage, we combine the results of the two models here. We calculate the average of the two uncertainties and take it as a weight coefficient and multiply it on the loss function. Since the calculated uncertainty fluctuates greatly, we normalize it in this way:

Uun=β×[0.5(Uun1+Uun2)+C],U_{un}=-\beta\times[0.5(U_{un1}+U_{un2})+C], (4)

where β\beta and CC are constants used for normalization. In order t to make the uncertainty change within the controllable range, we set β=0.7\beta=0.7, C=2C=2.

In supervised learning, the models should pay more attention to the pixels with higher uncertainty. But here in unsupervised learning, models should learn more from lower uncertainty pixels as there is no label in this stage. The estimated uncertainty guides models concentrating more on reliable (low uncertainty) predictions. We add uncertainty Uun1U_{un_{1}}, Uun2U_{un_{2}} estimated by two models together, and normalize it as UunU_{un}. The uncertainty-aware ensemble agreement loss we leverage here is Jensen-Shannon divergence:

agr=Uun𝔼Du[DJS(f(xiu;θ1)||f(xiu;θ2))].\displaystyle\mathcal{L}_{agr}=U_{un}\mathbb{E}_{D_{u}}[D_{JS}(f(x_{i}^{u};{\theta_{1}})||f(x_{i}^{u};{\theta_{2}}))]. (5)
Refer to caption
Figure 3: Illustration of unsupervised learning. Two networks output predictions based on the knowledge from each view. We leverage the ensemble agreement loss con\mathcal{L}_{con} to encourage the predictions of models to consistency with the guidance of the estimated uncertainty. The final uncertainty is the sum of the uncertainties generated by two models using Monte Carlo Dropout. Each model performs TT stochastic forward passes to get an uncertainty map.
Refer to captionRefer to caption
Refer to captionRefer to caption
Figure 4: Illustration of the DSC(%) and HD(mm) results of introducing uncertainty in different epochs in the supervised and unsupervised stages. Pictures (a), (b) represent leveraging uncertainty only in the unsupervised stage, and (c), (d) represent using uncertainty only in the supervised stage. The red broken lines in each picture represent voting results by two models, and the blue broken lines represent average results. Considering the performance both in DSC and HD, we finally choose to add uncertainty at 0 epoch of supervised stage, and at 20 epoch of unsupervised stage.

3.2.3 When to introduce uncertainty?

Introducing uncertainty into models can improve the performance of the network, but introducing it at unappropriate time will crush the entire framework.

In the supervised stage, theoretically, the models don’t need prior knowledge to avoid turning in the wrong direction. So the sooner uncertainty is introduced to guide network learning, the better performance we can get. We verify this by doing experiments of taking uncertainty in concern at different epoch. The results are shown in Fig. 4. Considering the performance in both DSC and HD, we leverage uncertainty from the beginning in the supervised stage.

In the unsupervised stage, the models’ prediction error for unlabeled images is large at the beginning of experiments. This will lead to a large fluctuation range of uncertainty, which is not conducive to allowing the model to focus on where it should learn more. So, we tested adding uncertainty to the unsupervised stage at different epoch, and the final results are shown in Fig. 4. And finally, we learned that introducing uncertainty in the unsupervised stage requires that models have a certain understanding of the target distribution.

Synthesize all factors, we introduce uncertainty to the supervised stage at epoch 0, and epoch 20 for the unsupervised stage in all experiments of our approach.

3.2.4 Adversarial learning

Having diversity between models in the ensemble is indispensable. In co-training, diversity is essential for avoiding the collapse of decision boundary and the models can learn from each other during training. The deep co-training method uses the approach proposed by Qiao et al. for image classification and augments the dataset with adversarial examples. The adversarial examples are generated from both labeled and unlabeled data.

One model uses adversarial examples teaching other models in the ensemble. In the original deep co-training method, the diversity loss is:

div(θ;D)=𝔼xD[(f1(x;θ1),f2(g1(x);θ2))\displaystyle\mathcal{L}_{div}(\theta;D)=\mathbb{E}_{x\in D}[\mathcal{H}(f^{1}(x;\theta^{1}),f^{2}(g^{1}(x);\theta^{2})) (6)
+(f2(x;θ2),f1(g2(x);θ1))],\displaystyle+\mathcal{H}(f^{2}(x;\theta^{2}),f^{1}(g^{2}(x);\theta^{1}))],

where (.)\mathcal{H}(.) refers to cross entropy and xx is an input image. The gi(x)g^{i}(x) is an adversarial example target on model fi(.;θi)f^{i}(.;\theta^{i}). So g1(x)g^{1}(x) is an adversarial example for model 1, and we can easily get that f1(x;θ1)f^{1}(x;\theta^{1}) \not= f1(g1(x);θ1)f^{1}(g^{1}(x);\theta^{1}). In the process of minimizing the first term of Eq. 6., it will make f1(x;θ1)f^{1}(x;\theta^{1}) approximately equal to f2(g1(x);θ2)f^{2}(g^{1}(x);\theta^{2}). Combining the above relations, we can obtain f1(g1(x);θ1)f^{1}(g^{1}(x);\theta^{1}) \neq f2(g1(x);θ2)f^{2}(g^{1}(x);\theta^{2}). Applying the same idea to model 2, we can conclude that the models have divergence in predicting adversarial examples.

Adversarial examples can be obtained by adding small perturbations to input images. As described in  [13], we follow use different schemes to generate these examples based on the source of the image xx. We apply the Virtual Adversarial Training (VAT) method when xx is an unlabeled image from DuD_{u}; and we use the Fast Gradient Sign Method (FGSM) when xx in labeled datasets.

Table 1: Comparison of semi-supervised method results on segmentation from ACDC dataset with lαl_{\alpha}=0.2. All the experiments are reimplemented by us.
Method DSC(%\%)
RV Myo LV Mean
Full 89.33(0.23)89.33(0.23) 88.07(0.53)88.07(0.53) 93.34(0.48)93.34(0.48) 90.25(0.48)90.25(0.48)
MT 76.70(7.07)76.70(7.07) 74.26(4.78)74.26(4.78) 84.73(1.24)84.73(1.24) 78.56(2.86)78.56(2.86)
UAMT 77.69(1.04)77.69(1.04) 78.88(1.21)78.88(1.21) 88.17(0.57)88.17(0.57) 81.58(0.32)81.58(0.32)
Part avg 73.89(1.90)73.89(1.90) 75.64(2.25)75.64(2.25) 86.95(0.85)86.95(0.85) 78.83(1.66)78.83(1.66)
vot 75.16(2.00)75.16(2.00) 77.61(2.39)77.61(2.39) 88.38(0.89)88.38(0.89) 80.39(1.72)80.39(1.72)
DCT avg 81.89(1.95)81.89(1.95) 83.32(0.78)83.32(0.78) 91.13(0.45)91.13(0.45) 85.45(0.68)85.45(0.68)
vot 82.45(2.17)82.45(2.17) 84.35(0.58)84.35(0.58) 91.68(0.62)91.68(0.62) 86.16(0.73)86.16(0.73)
Ours avg 85.91(0.12)85.91(0.12) 85.96(0.27)85.96(0.27) 92.34(0.26)92.34(0.26) 88.07(0.15)88.07(0.15)
vot 86.59(0.17) 86.65(0.35) 92.74(0.33) 88.45(0.12)
Method HD(mmmm)
RV Myo LV Mean
Full 3.77(0.15)3.77(0.15) 3.09(0.06)3.09(0.06) 2.40(0.10)2.40(0.10) 3.09(0.03)3.09(0.03)
MT 10.03(1.07)10.03(1.07) 10.47(3.31)10.47(3.31) 10.34(4.09)10.34(4.09) 10.28(2.25)10.28(2.25)
UAMT 8.05(0.93)8.05(0.93) 7.18(3.31)7.18(3.31) 6.30(2.18)6.30(2.18) 7.18(1.12)7.18(1.12)
Part avg 14.09(0.52)14.09(0.52) 12.14(0.80)12.14(0.80) 8.87(0.74)8.87(0.74) 11.70(0.42)11.70(0.42)
vot 7.49(0.23)7.49(0.23) 6.25(0.43)6.25(0.43) 4.57(0.19)4.57(0.19) 6.11(0.24)6.11(0.24)
DCT avg 8.16(0.93)8.16(0.93) 7.42(1.65)7.42(1.65) 5.38(0.12)5.38(0.12) 7.23(0.48)7.23(0.48)
vot 4.84(0.57) 4.06(0.19)4.06(0.19) 3.01(0.08)3.01(0.08) 3.97(0.22)3.97(0.22)
Ours avg 6.74(0.33)6.74(0.33) 6.34(0.48)6.34(0.48) 4.60(0.30)4.60(0.30) 5.89(0.25)5.89(0.25)
vot 4.93(1.30)4.93(1.30) 3.65(0.18) 2.68(0.08) 3.48(0.05)
Table 2: Comparison of semantic segmentation results with different label ratios in ACDC dataset.
Method DSC(%)
55 1010 2020 3030 4040 5050
Part 69.3969.39 76.3476.34 79.1279.12 88.0688.06 89.1189.11 89.6789.67
IND avg 73.2673.26 73.5973.59 81.1681.16 86.1686.16 86.6486.64 86.6786.67
vot 73.6573.65 75.6675.66 82.8282.82 87.3087.30 88.0488.04 88.6888.68
DCT avg 79.4579.45 85.6785.67 85.2485.24 87.2887.28 88.3788.37 88.1988.19
vot 80.4380.43 86.4086.40 86.3186.31 88.0288.02 89.0189.01 89.1089.10
Ours avg 82.5282.52 87.5887.58 87.9987.99 88.9388.93 89.3789.37 90.4090.40
vot 83.16 88.01 88.57 89.41 89.86 90.89
Method HD(mm)
5 10 20 30 40 50
Part 9.299.29 6.536.53 12.8712.87 4.014.01 3.443.44 3.293.29
IND avg 12.1412.14 16.1316.13 12.2812.28 7.977.97 7.687.68 6.946.94
vot 5.595.59 8.258.25 5.825.82 4.434.43 4.274.27 3.833.83
DCT avg 11.5311.53 6.866.86 7.897.89 6.066.06 5.735.73 5.935.93
vot 6.046.04 4.114.11 4.274.27 3.583.58 3.383.38 3.623.62
Ours avg 7.597.59 5.535.53 6.016.01 5.335.33 5.075.07 4.894.89
vot 4.70 3.42 3.54 3.09 3.15 2.78

4 Experiments and results

4.1 Dataset and metrics

We evaluated our method on three medical image segmentation benchmark publicly available datasets: Automated Cardiac Challenge (ACDC)[32], Spinal Cord Gray Matter Challenge (SCGM)[33], and Spleen sub-task dataset of the Medical Segmentation Decathlon Challenge[34].

ACDC dataset

: The University Hospital of Dijon created the ACDC dataset from real clinical exams. The ACDC dataset covers several well-defined pathologies with enough cases. It consists of 200 short-axis cine-MRI scans from 100 patients. This dataset is divided into 5 evenly distributed medical groups: patients without cardiac disease, myocardial infarction, hypertrophic cardiomyopathy, abnormal right ventricles, and dilated cardiomyopathy. There are four regions of interest in segmentation masks: right ventricle endocardium (RV), left ventricle endocardium (LV), left ventricle myocardium (Myo), and background. We leveraged 75 subjects (150 scans) for training and 25 subjects (50 scans) for testing. All short-axis slices within 3D-MRI scans were resized to 256 ×\times 256 as 2D images.

SCGM dataset

: The Spinal Cord Gray Matter Challenge was organized to test the different capabilities of various methods. The dataset in this challenge is a public-available collection of multi-vendors, multi-center MRI. All data was acquired at 4 different sites: University College London, Polytechnique Montreal, University of Zurich, and Vanderbilt University. It contains 80 healthy subjects (age range of 28.3 to 44.3 years) with 20 subjects from each center. The training set contains 40 labeled scans, each annotated slice-wise by 4 independent experts. The ground truth mask is obtained by majority voting. In our work, we leverage 30 images from the first center as a labeled dataset, and 465 images from all centers as an unlabeled dataset. The test set contains 264 labeled images from center 3 and center 4. The slices are center-cropped to 200 ×\times 200 pixels.

Spleen dataset

: Spleen dataset is one of the ten sub-tasks of the Medical Segmentation Decathlon Challenge[34]. It consists of patients undergoing chemotherapy treatment for liver metastases as a publicly available dataset. The dataset includes 61 portal venous phases CT scans (only 41 were given with ground truth). The ground truth segmentation was generated by a semi-automatic segmentation software and identified by an expert abdominal radiologist. Each slice obtained from the CT scans is resized to 256 ×\times 256 pixels. We split the dataset into labeled (4 patients), unlabeled (32 patients), and validation image subsets( 5 patients).

Dice similiarity coefficient(DSC)

: DSC measures the overlap between the predictions of the model SS and the ground truth GG:

DSC(S,G)=2|SG||S|+|G|.DSC(S,G)=\frac{2|S\cap G|}{|S|+|G|}. (7)
Hausdorff distance(HD)

: HD is a boundary distance metric which measures the largest distance (in mm) between a point in prediction SS and the closest point in the ground truth GG:

HD(S,G)=max{d(S,G),d(G,S)}.HD(S,G)=max\begin{Bmatrix}d(S,G),d(G,S)\end{Bmatrix}. (8)
Refer to caption
Figure 5: Examples of segmentation results for ACDC dataset with 20%\% label images. From left to right: Ground truth (GT), Fully supervised baseline (Full), Independent, Mean Teacher (MT), Uncertainty-aware Mean Teacher (UAMT), Deep co-training (DCT), and our Uncertainty-aware Deep co-training method (Ours).
Refer to caption
Figure 6: Examples of segmentation results for SCGM dataset using center 1 as training data. From left to right: Ground truth (GT), Fully supervised baseline (Full), Independent, Mean Teacher (MT), Uncertainty-aware Mean Teacher (UAMT), Deep co-training (DCT), and our Uncertainty-aware Deep co-training method (Ours).
Refer to caption
Figure 7: Examples of segmentation results for Spleen dataset with 20% label images. From left to right: Ground truth (GT), Fully supervised baseline (Full), Independent, Mean Teacher (MT), Uncertainty-aware Mean Teacher (UAMT), Deep co-training (DCT), and our Uncertainty-aware Deep co-training method (Ours).
Refer to caption
Figure 8: Uncertainty heatmap under different epochs in supervised learning.
Refer to caption
Figure 9: Uncertainty heatmap under different epochs in unsupervised learning.
Table 3: Comparison of semi-supervised method results on segmentation from SCGM dataset.
Method DSC(%\%) HD(mmmm)
Full 61.37(3.05)61.37(3.05) 32.27(13.52)32.27(13.52)
Mean Teacher 49.13(1.07)49.13(1.07) 16.72(2.25)16.72(2.25)
UAMT 39.82(2.28)39.82(2.28) 39.35(6.04)39.35(6.04)
Independent avg 48.09(0.65)48.09(0.65) 33.42(2.01)33.42(2.01)
vot 47.49(0.12)47.49(0.12) 37.23(4.10)37.23(4.10)
DCT avg 66.02(1.92)66.02(1.92) 18.56(2.21)18.56(2.21)
vot 68.35(1.97)68.35(1.97) 15.36(3.85)15.36(3.85)
Ours avg 73.19(0.25)73.19(0.25) 10.05(1.05)10.05(1.05)
vot 74.31(0.30) 9.01(0.73)
Table 4: Comparison of semi-supervised method results on segmentation from Spleen dataset.
Method DSC(%\%) HD(mmmm)
Full 94.56(0.19)94.56(0.19) 6.84(0.52)6.84(0.52)
Mean Teacher 82.25(1.45)82.25(1.45) 9.38(1.58)9.38(1.58)
UAMT 84.07(1.02)84.07(1.02) 15.48(4.46)15.48(4.46)
Independent avg 91.99(0.35)91.99(0.35) 12.44(1.36)12.44(1.36)
vot 92.85(0.29)92.85(0.29) 10.75(0.32)10.75(0.32)
DCT avg 93.08(0.60)93.08(0.60) 8.88(1.73)8.88(1.73)
vot 94.09(0.58)94.09(0.58) 6.86(0.61)6.86(0.61)
Ours avg 94.55(0.19)94.55(0.19) 7.24(0.14)7.24(0.14)
vot 95.45(0.24) 5.57(0.21)
Table 5: Comparison of different label ratios in Spleen dataset.
Method DSC
5%5\% 10%10\%
Independent avg 84.51(0.97)84.51(0.97) 88.18(0.42)88.18(0.42)
vot 86.94(0.61)86.94(0.61) 89.79(0.73)89.79(0.73)
DCT avg 86.77(2.63)86.77(2.63) 88.56(0.58)88.56(0.58)
vot 87.91(2.63)87.91(2.63) 90.66(0.27)90.66(0.27)
Ours avg 87.76(1.59)87.76(1.59) 90.06(0.51)90.06(0.51)
vot 89.02(1.05) 91.94(0.77)
Table 6: Abalation analysis of our method.
Method DSC(%) HD(mm)
RV Myo LV Mean RV Myo LV Mean
DCT-seg avg 81.76 82.21 91.75 85.24 9.31 9.16 5.21 7.89
vot 82.84 83.55 92.53 86.31 5.64 4.26 2.90 4.27
Supervised stage via uncertainty avg 85.79 85.75 92.11 87.88 6.66 6.60 4.59 5.95
vot 86.51 86.50 92.17 88.39 3.93 3.52 2.76 3.40
Unsupervised stage via uncertainty avg 84.55 84.73 92.26 87.18 7.42 5.75 4.50 5.89
vot 85.29 85.37 92.45 87.70 4.42 3.41 2.72 3.52

4.2 Implementation Details

For the sake of speed and accuracy, we leverage the well-known lightweight E-net[31] as our basic segmentation network. This architecture is one of the most popular models for segmentation. To adapt the E-Net as a Bayesian network to estimate the uncertainty, one dropout layer with a dropout rate of 0.50.5 is added between the encoder and decoder of the E-Net. All the datasets are applied random rotation, random crop, and flip as augmentation strategies. We set ϵ\epsilon to 0.03 in FSGM and 10 in VAT. The learning rate decreased with a factor of 10 every 90 epochs, and it is 10 every 100 epochs in SCGM experiments. To describe the level of supervision, we vary the ratio lal_{a} (0 \leq lal_{a} \leq 1) of labeled samples in our experiments.

For all experiments, the setting of λ\lambda is with a dynamic strategy. We used a ramp-up strategy followed the Gaussian ramp-up curve. And for all the hyperparameter λ\lambda settings, we follow  [13]. The framework is implemented with Pytorch library, trained on one NVIDIA 2080Ti GPU.

4.3 Comparison with semi-supervised methods

Our uncertainty-aware co-training method is compared against several recent state-of-art semi-supervised methods in the medical domain:

Mean Teacher

: Mean Teacher (MT) is a method using multiple deep CNNs for semi-supervised segmentation. And it is an effective approach that averages model weights instead of predictions.

Uncertainty-aware Mean Teacher

: Uncertainty-aware Mean Teacher (UAMT) applies uncertainty estimation to MT.

Deep-Co-trianing

: Deep-Co-training (CT) shares information between simultaneously trained models, while preserving their diversity.

We also compare our method with the full supervised method of the same amount of labeled images as the semi-supervised method, referred to as ”Part”. ”IND” means the performance of individually-trained models (Independent). We report the performance of a fully supervised baseline (”Full”) which means training a single model with all available datasets.

For these baselines, we follow the same learning rate decay, weight scheduler, data augmentation setting, and optimization as for our method. For MT and UAMT, we set EMA (Exponential Moving Average) parameters α\alpha to 0.99. We report both the average performance of individual models (”avg”) and the performance of combining the predictions of all models using a voting strategy (”vot”) called ensemble soft voting. This strategy leads to a higher accuracy than the prediction of individual models for both DSC and HD. In order to avoid contingency, we compute the average and standard deviation over three runs with different random seeds.

All the compared baseline results reported in our paper are reimplemented by us.

4.4 Experimental results

4.4.1 ACDC dataset

Our uncertainty-aware method is first evaluated on the ACDC dataset with lal_{a} = 0.2. Tab. 2. presents the quantitative results. Compared with other methods, our approach achieves higher overall performance on the test dataset in terms of DSC and HD. Especially, our approach achieves improvement over co-training with an overall performance increase of 2.29%\% in DSC and 0.49mmmm in HD, and the standard deviation of multiple trials (three runs) of the results of our approach is the smallest. With only 20%\% labeled images, we are only 1.8%\% in DSC and 0.39mmmm in HD behind the 100%\% label ratio supervision. Some examples of results from the test dataset are shown in Fig. 5. We can see our method gives contours closer to ground truth (GT), with more accurate segmentation in details between different regions.

Also, we evaluate how label ratios impact results in a dual-view setting. Tab. 3. shows the results for different labeled data ratios: 5%\%, 10%\%, 20%\%, 30%\%, 40%\%, 50%\%. We can easily find that as the label ratio increase, mean DSC values increase sharply, and mean HD values decrease. In all cases, our approach leads to a better performance in DSC and HD than training models separately and deep co-training method. Our approach has more obvious advantages with a low label ratio. With 5%\% label ratio, our method outperforms Deep co-training 2.73%\% in DSC and 1.34mmmm in HD.

4.4.2 SCGM dataset

We evaluate our method on the task of segmenting spinal cord grey matter in images from the SCGM dataset. The SCGM dataset is from four different clinical centers, so different parameters are applied in collecting the MRI images. We only used a few labeled images (i.e., only 30 images from one center), and test sets are from the other centers. The labeling ratio is about 6.5%\%. Due to the different data sources, it is more difficult to extract the image features of the samples. This also leads to low segmentation accuracy of semi-supervised methods in this task. The use of uncertainty solves this problem well, models can learn effective semantic features from different centers according to the uncertainty.

The results are summarized in Tab. 4. Our approach gives a mean DSC of 5.96%\% and HD of 6.35mmmm better than the best baseline (deep co-training). And our method gives a mean DSC of 12.94%\% and HD of 23.26mmmm better than the performance of the fully supervised baseline. Fig. 6. shows the segmentation results on the test dataset. In some difficult-to-recognize images, fully supervised training cannot even segment the lesion area, but our approach can complete the segmentation task to a certain extent.

4.4.3 Spleen dataset

We also validate the effectiveness of our uncertainty-aware deep co-training method on the task of segmenting spinal cord grey matter in images from the Spleen dataset. We repeated our experiments on the Spleen dataset consisting of 2D slices of CT scans resized to a resolution of 256 ×\times 256 pixels. Tab. 5. summarizes the experimental results. We see that the Deep co-training method’s performance is almost the same as the fully supervised baseline. In case of better stability (standard deviation) of the results, the accuracy has been improved and surpassed full supervision after taking uncertainty into concern. Our method improves accuracy over DCT-seg from 94.09%\% to 95.45%\% and HD from 6.86mmmm to 5.57mmmm. Examples of segmentation results obtained by tested methods are given in Fig. 7.

Semi-supervised learning pursues using less labeled images to achieve better segmentation accuracy. We show the performance of our approach, DCT-seg, and individually trained models(independent) on smaller label ratios: 5%\% and 10%\%. Tab.6. gives the mean DSC with standard deviation. Our approach is still in the leading position among all semi-supervised methods.

4.4.4 Abalation analysis of our method

We do ablation studies to prove that our method works in both supervised and unsupervised stages on the ACDC dataset. The results are shown on Tab.6.

In supervised stage: As mentioned before, the best time to introduce uncertainty to the full supervision phase is from the beginning. Uncertainty used here allows models in the network to obtain characteristics of the target area better under full supervision. We can see from Fig. 8. that the main effect of uncertainty in the full supervision stage is in the first 30 epochs. We can tell from the label that the uncertainty maps focus more on the lesion area. After adding uncertainty to the full supervision stage, the method can achieve improvement over co-training with an overall all performance increase of 2.23%\% in DSC and 0.57mmmm in HD.

In unsupervised stage: The best time at this stage is to leverage uncertainty after models have a certain understanding of the lesion area’s semantic features. After experiments per 10 epochs from 0, it is finally determined that the uncertainty will be leveraged after the 20th epoch. So, we visualize the uncertainty map from 0 epochs to the last in Fig. 9. We can see that the uncertainty maps are increasingly focusing on the lesion areas. Obviously, uncertainty always plays a role throughout the experiment. As the training progresses, the color of the lesion areas of the uncertainty maps becomes lighter, which means the predictions of the two models for the same unlabeled image are gradually approaching, but there is still a certain difference. Only introduce uncertainty to the unsupervised stage, we can achieve an improvement over co-training with an overall all performance increase of 1.39%\% in DSC and 0.75mmmm in HD.

5 Conclusion

In this paper, we propose a novel uncertainty-aware deep co-training method for three medical image segmentation tasks. Our approach uses uncertainty in both the supervised learning stage and the unsupervised learning stage in the deep co-training method. We use uncertainty obtained from Monte Carlo Sampling to guide the training process purposely. We validate our method on three challenging medical image datasets. The comparison with other semi-supervised methods confirms the effectiveness of our approach. Our uncertainty-aware co-training method achieves the performance 1.88%\% away from supervision with only 20%\% label on ACDC dataset, and an increase of 12.94%\% in terms of DSC over fully supervised method on SCGM dataset. And for the Spleen dataset, we get 0.89%\% more than the fully supervised approach in DSC. In future work, we will investigate the effect of different uncertainty estimation manners and apply our approach to other semi-supervised medical image segmentation tasks.

References

  • [1] M. Z. Khan, M. K. Gajendran, Y. Lee, M. A. Khan, Deep neural architectures for medical image semantic segmentation, IEEE Access (2021).
  • [2] H. Noh, S. Hong, B. Han, Learning deconvolution network for semantic segmentation, in: Proceedings of the IEEE international conference on computer vision, 2015, pp. 1520–1528.
  • [3] R. Yang, Y. Yu, Artificial convolutional neural network in object detection and semantic segmentation for medical imaging analysis, Frontiers in Oncology 11 (2021) 573.
  • [4] S. A. Taghanaki, K. Abhishek, J. P. Cohen, J. Cohen-Adad, G. Hamarneh, Deep semantic segmentation of natural and medical images: a review, Artificial Intelligence Review 54 (1) (2021) 137–178.
  • [5] L. Schmarje, M. Santarossa, S.-M. Schröder, R. Koch, A survey on semi-, self-and unsupervised learning for image classification, IEEE Access (2021).
  • [6] V. Cheplygina, M. de Bruijne, J. P. Pluim, Not-so-supervised: a survey of semi-supervised, multi-instance, and transfer learning in medical image analysis, Medical image analysis 54 (2019) 280–296.
  • [7] Y. Meng, H. Zhang, Y. Zhao, X. Yang, X. Qian, X. Huang, Y. Zheng, Spatial uncertainty-aware semi-supervised crowd counting, in: Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 15549–15559.
  • [8] Y. Li, X. Chao, Semi-supervised few-shot learning approach for plant diseases recognition, Plant Methods 17 (1) (2021) 1–10.
  • [9] C. Baur, S. Albarqouni, N. Navab, Semi-supervised deep learning for fully convolutional networks, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2017, pp. 311–319.
  • [10] X. Li, L. Yu, H. Chen, C.-W. Fu, P.-A. Heng, Semi-supervised skin lesion segmentation via transformation consistent self-ensembling model, in: Proceedings of the British Machine Vision Conference (BMVC), 2018.
  • [11] D.-H. Lee, et al., Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks, in: Workshop on challenges in representation learning, ICML, Vol. 3, 2013, p. 896.
  • [12] W.-C. Hung, Y.-H. Tsai, Y.-T. Liou, Y.-Y. Lin, M.-H. Yang, Adversarial learning for semi-supervised semantic segmentation, in: Proceedings of the British Machine Vision Conference (BMVC), 2018.
  • [13] J. Peng, G. Estrada, M. Pedersoli, C. Desrosiers, Deep co-training for semi-supervised image segmentation, Pattern Recognition 107 (2020) 107269.
  • [14] A. Kendall, Y. Gal, What uncertainties do we need in bayesian deep learning for computer vision?, in: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Curran Associates Inc., Red Hook, NY, USA, 2017, p. 5580–5590.
  • [15] I. J. Goodfellow, J. Shlens, C. Szegedy, Explaining and harnessing adversarial examples, arXiv preprint arXiv:1412.6572 (2014).
  • [16] X. J. Zhu, Semi-supervised learning literature survey (2005).
  • [17] W. Bai, O. Oktay, M. Sinclair, H. Suzuki, M. Rajchl, G. Tarroni, B. Glocker, A. King, P. M. Matthews, D. Rueckert, Semi-supervised learning for network-based cardiac mr image segmentation, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2017, pp. 253–260.
  • [18] X. L. C.-W. F. P.-A. H. Lequan Yu, Shujun Wang, Uncertainty-aware self-ensembling model for semi-supervised 3d left atrium segmentation, in: Medical Image Computing and Computer Assisted Intervention Society, 2019.
  • [19] K. Fang, W.-J. Li, Dmnet: Difference minimization network for semi-supervised segmentation in medical images, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2020, pp. 532–541.
  • [20] W.-C. Hung, Y.-H. Tsai, Y.-T. Liou, Y.-Y. Lin, M.-H. Yang, Adversarial learning for semi-supervised semantic segmentation, arXiv preprint arXiv:1802.07934 (2018).
  • [21] Y. Wu, M. Xu, Z. Ge, J. Cai, L. Zhang, Semi-supervised left atrium segmentation with mutual consistency training, arXiv preprint arXiv:2103.02911 (2021).
  • [22] X. Zhang, Z. Cui, C. Chen, J. Wei, J. Lou, W. Hu, H. Zhang, T. Zhou, F. Shi, D. Shen, Confidence-aware cascaded network for fetal brain segmentation on mr images, in: International Conference on Medical Image Computing and Computer-Assisted Intervention, Springer, 2021, pp. 584–593.
  • [23] W. Edeling, H. Arabnejad, R. Sinclair, D. Suleimenova, K. Gopalakrishnan, B. Bosak, D. Groen, I. Mahmood, D. Crommelin, P. V. Coveney, The impact of uncertainty on predictions of the covidsim epidemiological code, Nature Computational Science 1 (2) (2021) 128–135.
  • [24] M. Abdar, F. Pourpanah, S. Hussain, D. Rezazadegan, L. Liu, M. Ghavamzadeh, P. Fieguth, X. Cao, A. Khosravi, U. R. Acharya, et al., A review of uncertainty quantification in deep learning: Techniques, applications and challenges, Information Fusion (2021).
  • [25] C. E. Papadopoulos, H. Yeung, Uncertainty estimation and monte carlo simulation method, Flow Measurement and Instrumentation 12 (4) (2001) 291–298.
  • [26] B. Lakshminarayanan, A. Pritzel, C. Blundell, Simple and scalable predictive uncertainty estimation using deep ensembles (2017). arXiv:1612.01474.
  • [27] A. Atanov, A. Ashukha, D. Molchanov, K. Neklyudov, D. Vetrov, Uncertainty estimation via stochastic batch normalization, arXiv preprint arXiv:1802.04893 (2018).
  • [28] C. Louizos, M. Welling, Multiplicative normalizing flows for variational bayesian neural networks, in: International Conference on Machine Learning, PMLR, 2017, pp. 2218–2227.
  • [29] U. Upadhyay, V. P. Sudarshan, S. P. Awate, Uncertainty-aware gan with adaptive loss for robust mri image enhancement, in: Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, 2021, pp. 3255–3264.
  • [30] X. Yan, S. Hu, Y. Mao, Y. Ye, H. Yu, Deep multi-view learning methods: a review, Neurocomputing (2021).
  • [31] A. Paszke, A. Chaurasia, S. Kim, E. Culurciello, Enet: A deep neural network architecture for real-time semantic segmentation (2016). arXiv:1606.02147.
  • [32] O. Bernard, A. Lalande, C. Zotti, F. Cervenansky, X. Yang, P.-A. Heng, I. Cetin, K. Lekadir, O. Camara, M. A. Gonzalez Ballester, G. Sanroma, S. Napel, S. Petersen, G. Tziritas, E. Grinias, M. Khened, V. A. Kollerathu, G. Krishnamurthi, M.-M. Rohé, X. Pennec, M. Sermesant, F. Isensee, P. Jäger, K. H. Maier-Hein, P. M. Full, I. Wolf, S. Engelhardt, C. F. Baumgartner, L. M. Koch, J. M. Wolterink, I. Išgum, Y. Jang, Y. Hong, J. Patravali, S. Jain, O. Humbert, P.-M. Jodoin, Deep learning techniques for automatic mri cardiac multi-structures segmentation and diagnosis: Is the problem solved?, IEEE Transactions on Medical Imaging 37 (11) (2018) 2514–2525. doi:10.1109/TMI.2018.2837502.
  • [33] F. Prados, J. Ashburner, C. Blaiotta, T. Brosch, J. Carballido-Gamio, M. J. Cardoso, B. N. Conrad, E. Datta, G. Dávid, B. De Leener, et al., Spinal cord grey matter segmentation challenge, Neuroimage 152 (2017) 312–329.
  • [34] A. L. Simpson, M. Antonelli, S. Bakas, M. Bilello, K. Farahani, B. van Ginneken, A. Kopp-Schneider, B. A. Landman, G. Litjens, B. Menze, O. Ronneberger, R. M. Summers, P. Bilic, P. F. Christ, R. K. G. Do, M. Gollub, J. Golia-Pernicka, S. H. Heckers, W. R. Jarnagin, M. K. McHugo, S. Napel, E. Vorontsov, L. Maier-Hein, M. J. Cardoso, A large annotated medical image dataset for the development and evaluation of segmentation algorithms (2019). arXiv:1902.09063.