ATSO: Asynchronous Teacher-Student Optimization for Semi-Supervised Medical Image Segmentation
Abstract
In medical image analysis, semi-supervised learning is an effective method to extract knowledge from a small amount of labeled data and a large amount of unlabeled data. This paper focuses on a popular pipeline known as self-learning, and points out a weakness named lazy learning that refers to the difficulty for a model to learn from the pseudo labels generated by itself. To alleviate this issue, we propose ATSO, an asynchronous version of teacher-student optimization. ATSO partitions the unlabeled data into two subsets and alternately uses one subset to fine-tune the model and updates the label on the other subset. We evaluate ATSO on two popular medical image segmentation datasets and show its superior performance in various semi-supervised settings. With slight modification, ATSO transfers well to natural image segmentation for autonomous driving data.
1 Introduction
Semantic segmentation plays an important role in medical image analysis. Recently, the fast development of deep learning [10] provides a powerful tool for dense image prediction [2, 13], but for many scenarios of medical image analysis, data annotation is often expensive and thus difficult to acquire. This falls into the area of semi-supervised learning which focuses on learning from both labeled data and unlabeled data, but the labeled part is often much smaller. An effective pipeline is known as self-learning, in which an initial model is trained on the labeled part (training set) and fine-tuned on the unlabeled part (reference set) with the pseudo labels generated by itself.
We refer to this pipeline as teacher-student optimization, a variant of knowledge distillation [7] that has straightforward applications on medical image analysis [36]. However, we notice a factor that harms the efficiency of utilizing the unlabeled data. This is because in the continual learning procedure, the teacher and student models tend to have increasing similarity to each other. Consequently, the new supervision that the student model obtains from the pseudo labels becomes weaker and weaker, and the learning process quickly arrives at a plateau that is below satisfaction. This phenomenon is called lazy learning which, quantitatively, often reflects in that the quality of pseudo labels does not get improved during the learning process – in other words, the accuracy on the reference set stops growing but the model itself does not know.
As the key observation of this paper, we point out that the essence of lazy learning is that the self-learning process is gradually pushing the teacher and student models, as a whole, towards a local optimum. In particular, the teacher model stores knowledge in the pseudo labels for the student model to learn; once a system error appears, it is likely to persist throughout the iteration. Therefore, inaccuracy accumulates and finally results in the unsatisfying quality of pseudo labels. To alleviate this problem, we present the asynchronous teacher-student optimization (ATSO) algorithm that proposes two modifications, both of which are designed to break up the chain of ‘error propagation’. First, we switch off continual learning and start each generation from the same initialized model. Second, we prevent using the pseudo labels generated by a teacher model to supervise its direct student. The second modification involves partitioning the reference set into two subsets. In each round of teacher-student optimization, we generate the pseudo labels on any subset based on a teacher model that was not trained on the same set of data. As we shall see in experiments, this strategy can largely improve the accuracy on the reference set and, consequently, on the unseen test set.
We perform experiments on the NIH and MSD datasets for pancreas segmentation from CT scans. Besides investigating semi-supervised learning within NIH dataset ( or training data are labeled while others are not), we also investigate the issue of transfer learning (one dataset is fully annotated but the other is unlabeled, yet the goal is to achieve high accuracy on the second test set). In both cases, ATSO consistently outperforms the synchronous counterpart, as well as other semi-supervised learning methods by a significant margin. We also evaluate ATSO on Cityscapes and Mapillary, two autonomous driving datasets, showing its superiority in the domain of natural images.
2 Asynchronous Teacher-Student Optimization
2.1 Problem Setting, Fully-Supervised and Semi-Supervised Baselines
We study the problem of medical image segmentation in abdominal CT scans, and make use of 2D-based methods to process volumetric data. Each volume is partitioned into a set of slices along three directions, which are named the coronal, sagittal, and axial views in the context of medical images. Each slice is then sent into a 2D network for segmentation, and the output is stacked into a 3D volume as the final prediction. From this point of view, the key problem is a standard semantic segmentation task and the evaluation protocol is called the Dice-Sørensen coefficient (DSC), with the formula being , where and are the set of target voxels in the prediction and ground-truth, respectively. We choose the 2D segmentation baseline to be RSTN [33], an open-sourced, coarse-to-fine framework that achieved good performance in a few medical image segmentation tasks. Let us denote the network as , in which and denote the 2D input and output of the deep network parameterized by weights .
Although RSTN as well as other algorithms [8, 29, 17] achieved satisfying performance under full supervision, the performance can degenerate dramatically in the scenarios of limited training data. For example, a fully-supervised RSTN ( training samples) achieves an average accuracy of over on the NIH dataset, however, if only of training data ( training samples) are preserved, the accuracy quickly drops to nearly . This raises an important setting named semi-supervised learning, in which a large portion of training data do not have labels but we need to learn as much knowledge as possible from them. As a formal definition, the training set is partitioned into two parts, namely, the supervised (labeled) set and the reference (unlabeled) set . Most often, we have . Also, there is a testing set, , which is invisible to the algorithm.
A simple and effective pipeline for semi-supervised learning is named self-learning. It starts with an initial model, denoted by , which is trained under supervised learning on . gets updated for a total of rounds. In the -th ( round (a.k.a., generation)), the reference subset, , is sent into the old model (often referred to as the teacher model), and the prediction is named the pseudo label in the current round. The training process of the student model, , then follows a regular supervised learning procedure on both and , with the supervision on coming from the pseudo labels. The trained student model of the current round is used as the teacher model of the next round and the iteration continues till the end.
2.2 Lazy Learning: the Devil in Self-Learning
We start with training the model using of data labeled and the remaining unlabeled on the NIH dataset. The detailed settings are elaborated in Section 4.1. After generations, the self-learning pipeline achieves an accuracy of on the test set, claiming a significant improvement over that is reported by the model trained only on the labeled data. However, this number is still far below when the full labeled training set is available. That is to say, self-learning indeed extracts knowledge from the reference set, but the efficiency is far below satisfaction.
To understand the reason for this, we investigate the accuracy of the reference set, which is expected to grow with the learning procedure. However, we find that the accuracy quickly arrives at a plateau. After the th (the initial training stage with only labeled data), st, nd, and rd generations, the reference set accuracy is , , , and , respectively. Compared to the scenario that training data have been labeled, the accuracy on the same subset of training data is . In other words, the training procedure has entered a ‘trap’ that the reference set stops at a low accuracy (what is worse, the accuracy starts to drop in the rd generation), but the algorithm does not know because the ground-truth is missing.
From the viewpoint of optimization, this phenomenon can be explained as a local optimum of the self-learning system. Let be a training sample in the reference set, , is the ground-truth label, and is the predicted output by the teacher model. We assume that where is the prediction error. When is labeled, i.e., is known, follows a zero-mean distribution because the optimization goal is to minimize on the training set. However, in the scenario of self-learning, remains unknown and thus may follow a non-zero-mean distribution. Since is used as the pseudo label, such noise can persist across in the student model. What makes things worse, each generation of the teacher-student optimization can introduce new noise and the noise can accumulate on the reference set. We use the name of lazy learning to refer to the behavior that the student model is unable to identify and eliminate the noise of the teacher model.

2.3 Asynchronous Optimization: Escaping from the Trap
To alleviate lazy learning, the key is to escape from the optimization trap, i.e., breaking up the chain that the noise, , is propagated from the teacher to the student. This is done from two aspects, both of which focus on weakening the relationship between the teacher and student models.
First, we prevent using the setting of continual learning that initializes the student model from the teacher model and performs fine-tuning, but starts training from the very beginning. To reveal the advantage, we notice that the optimization goal is to minimize , but since is unknown, we use to estimate it. The error of this estimation is bounded by :
(1) |
That being said, provided that and are fixed. Hence, when is initialized independently from the teacher model, is increased and thus the fraction that occupies in the loss function is smaller. This effectively prevents the training procedure from being dominated by .
Second, we prevent using the same set of reference data continuously, i.e., always using the pseudo labels generated by a teacher model to supervise its direct student. To verify this assumption, we first notice that in the aforementioned self-learning procedure, the accuracy gained on the test set is much higher than that on the reference set, e.g., test accuracy gain vs. reference accuracy gain after self-learning generations. To make things clearer, we partition the reference set into two parts, and perform self-learning on different combinations of reference data. Results of two hard examples are shown in Figure 1. Interestingly, the segmentation accuracy is significantly improved when the example is not contained in the reference set. This aligns with the observation that the test accuracy is higher than the reference accuracy. Back to the optimization perspective, the noise on the reference set does not transfer to the data that the current generation does not use for self-learning. This inspires us to partition the reference set into two subsets, denoted by . In each generation, we generate the pseudo labels on using the model that was just self-trained on , and vice versa. After the last generation asynchronous update of both subsets, we combine the pseudo labels of both subsets into the complete reference set, based on which the final model is trained.
Integrating the above two aspects obtains the asynchronous teacher-student optimization (ATSO) algorithm that is described in Algorithm 1. Of course, there exist other solutions that can alleviate lazy learning, yet our solution is simple and effective. Moreover, ATSO also has the option of dividing the reference set into more subsets, but this can slow down the training procedure. In practice, we find that using two folds of reference data performs well for semi-supervised learning.
3 Related Work
Medical image analysis becomes increasingly important in assisting human doctors in clinics, in which semantic segmentation a fundamental basis of many tasks. This paper focuses on CT-scanned images, each of which is a 3D volume that corresponds to a specified region of the human body. Recently, with the fast development of deep neural networks [9, 22, 6], researchers developed effective algorithms for natural image segmentation [13, 2], and these techniques quickly propagated to the area of medical images [19, 15]. One of the major differences between segmenting natural and medical images lies in the dimensionality of input data. Medical scans often in 3D, but existing methods are often incapable of processing these data entirely. There are two popular pipelines, namely, 2D-based and 3D-based solutions, both of which sample data from the entire volume, process in the sampled region, and finally summarize all predictions into the original size. Differently, 2D-based pipelines [19, 21, 37, 33, 14] sampled 2D slices, while 3D-based pipelines [3, 15, 39] sample 3D sub-volumes. Both solutions have their advantages and disadvantages [25], and researchers tried to integrate them into one framework to absorb benefits from both of them [12, 16, 29].
Semi-supervised learning lies between supervised and unsupervised learning, which assumes that a small fraction of data are labeled, while the remaining part are unlabeled but closely related to the labeled subset [1, 38]. Researchers designed some generalized frameworks including self-learning [20], multi-view learning [24, 30], co-training [1], etc. The idea of self-learning is to use an initial model trained on labeled data to predict the labels on unlabeled data, so that these labels, though less accurate, can be used for training an updated, more powerful model [20]. This is related to knowledge distillation [7] and teacher-student optimization [5], but since unlabeled data was introduced, it is crucial to maximally improve the quality of the predicted label [26, 31]. On the other hand, both co-training and multi-view learning aim to use the consistency within the task itself to assist learning. Differently, co-training often assumed that different models should produce the same output on the same data [18], but multi-view learning assumed that the same model should produce the same result on various views of the same data [24]. Sometimes, these assumptions were combined into one framework [23, 18]. Semi-supervised learning is of great interest to the researchers of medical image analysis, mainly because accurate annotations are often difficult to acquire. There exists large-scale datasets with inaccurate [27] and/or partial data annotations [35], and researchers also developed practical semi-supervised algorithms for learning from these data [36, 28].
ATSO aims at improving the quality of ‘pseudo labels’ in the self-training pipeline of semi-supervised learning. The key principle is to enlarge the difference between the teacher and student signals so that the student model learns non-trivial knowledge from the teacher model. A similar idea was presented by a recent work [32] which studied fully-supervised learning tasks. Differently, [32] facilitates the difference by manipulating learning rates, while ATSO by isolating reference data between iterations.
ATSO is also related to other knowledge distillation approaches which trained a few models simultaneously so that each model can be used to supervise others. Examples include deep mutual learning [34] and deep co-training [18]. In particular, deep co-training [18] added an adversarial loss term to enlarge the gain between teacher and student models. Differently, we train two models individually on two subsets of the reference set, which naturally guarantees diversity and enjoys the ability of being parallelized when the number of individually-optimized models is large.
4 Semi-Supervised Learning Experiments
4.1 Datasets and Settings
We evaluate ATSO on the NIH dataset [21] for pancreas segmentation. It contains normal CT scans, each of which is a 3D volume of voxels, where is the length of the long axis. We follow the prior work [37, 33] to partition each dataset into four folds and use the first three folds ( cases) as the training data. For semi-supervised learning, we follow the prior work [36, 28] to use a small portion ( or ) of labeled training data and leave the remaining part to the reference set. We report the average DSC over all test cases.
The configuration of the deep networks follows that in the original RSTN paper [33]. In the training stage, we optimize three individual networks for segmentation along with the coronal, sagittal and axial views, respectively. In the inference stage, either on the reference set or the test set, predictions from these three views are fused into the final segmentation by majority voting. Please refer to the original paper for further technical details.
4.2 Quantitative and Qualitative Results
We first investigate semi-supervised segmentation on the NIH dataset with of the training set ( cases) labeled. The naive baseline, by only using the labeled data for training, reports a test accuracy which is far lower than the upper-bound, , when the labels of all training data are available. In what follows, we gradually add the key components of ATSO into the baseline and show how these components push the training procedure towards higher accuracy.
Breaking up Continual Learning Brings Significant Gain
We first compare the options with and without continual learning, which we refer to as the self-learning baseline and synchronous teacher-student optimization (STSO), respectively. The former uses the last snapshot of the teacher model to initialize the student model, while the latter trains the student model from scratch. To save computational costs, we define the scratch model as the snapshot after the first two sections of RSTN, after which a randomly initialized transformation module is inserted that injects sufficient perturbations to break up the effect of continual learning.
Results are summarized in Table 1. The difference between the self-learning baseline and STSO is significant. After two generations, the self-learning baseline achieves a accuracy on the testing set and on the reference set. Starting from the third generation, these numbers start to drop, demonstrating that lazy learning has obstructed the model from obtaining useful information. The best accuracy of STSO, , is obtained after generations. That is to say, simply switching off continual learning leads to a non-trivial improvement and, as we shall see in Table 2, surpasses all the competitors.
Asynchronous Optimization Further Improves Accuracy
Generation | Self-learning | STSO | ATSO | |||
---|---|---|---|---|---|---|
@ | @ | @ | @ | @ | @ | |
G0 | ||||||
G1 | ||||||
G2 | ||||||
G3 | ||||||
G4 | ||||||
G5 |
Next, we study the difference between STSO and ATSO. Results are summarized in Table 1. ATSO improves segmentation accuracy on both the reference and test sets. Detailed results on the reference subsets during iteration are provided in Appendix A. Interestingly, ATSO enjoys faster growth in both numbers: after only two generations, the test accuracy has increased to over , claiming a nearly advantage over the corresponding number of STSO. After five generations, ATSO still enjoys a advantage over STSO. That being said, ATSO has a broad range of applications in the scenario of limited computational resource for model training.
More importantly, these results indicate that the lazy learning phenomenon indeed persists during the entire training process. Therefore, by not generating pseudo labels on the reference set that was just used, the algorithm can escape from the optimization trap. We notice that compared to STSO that always refers to the entire reference set, ATSO trains each model using only half of the reference data, which has a natural deficit but still achieves higher accuracy. We expect ATSO to have a larger advantage when the amount of unlabeled data becomes larger.
Comparison to the State-of-the-Arts, and Visualization
In Table 2, we compare ATSO against state-of-the-art approaches, and show that ATSO outperforms all of them. In particular, ASTO surpasses [28] by more than in both scenarios that and labeled data have been used. Note that [28] is a recently published method which involved uncertainty in multi-view learning – in comparison, our solution is easier and more effective. In addition, being simple and easily implemented, ATSO can be combined with other training strategies, e.g., adversarial training [18] or uncertainty evaluation [28], towards better performance.
Figure 2 shows some typical examples of how segmentation errors are fixed with semi-supervised learning. When the labeled set is small, it is very likely that the labeled training set does cover sufficient situations, causing some failure cases in the reference set. In the self-learning baseline or even STSO, it is relatively difficult for the model to fix these errors during iteration, and the persisted errors can hinder the ability of the student model. In comparison, ATSO offers extra opportunities to jump out of the current distribution and thus get rid of the failure case. Hence, the efficiency of utilizing unlabeled training data is improved.
Method | Backbone | D | D | Method | Backbone | D | D |
---|---|---|---|---|---|---|---|
DMPCT [36] | 2D ResNet-101 | UMCT [28] (v fusion) | 3D ResNet-18 | ||||
DCT [18] (v) | 3D ResNet-18 | UMCT [28] (v fusion) | 3D ResNet-18 | ||||
TCSE [11] | 3D ResNet-18 | Self-Learning (ours) | 2D FCN8s | ||||
UMCT [28] (v) | 3D ResNet-18 | STSO (ours) | 2D FCN8s | ||||
UMCT [28] (v) | 3D ResNet-18 | ATSO (ours) | 2D FCN8s |

5 Transfer Learning Experiments
Beyond semi-supervised learning, we consider another scenario that has a broad range of applications, i.e., transferring the segmentation model trained on one dataset to another that does not have labels at all. For this purpose, we start with medical image segmentation and later generalize to natural images where a larger number of semantic classes are present.
5.1 Transfer Segmentation from NIH to MSD
The pancreatic tumor subset of the MSD dataset (http://medicaldecathlon.com/) has abnormal CT scans. Each sample contains the annotation of the pancreas and the tumor. Since we train the base model on the NIH dataset (normal pancreas), we only evaluate the pancreas segmentation accuracy on the MSD data. Transferring from NIH to MSD (the NIH training data are labeled while the MSD training data are not) is a challenging task since the distributions are very different between these two datasets: the scanners are different, and the MSD data contain abnormality (pancreatic tumor) while the NIH data do not. In this scenario, the key is to extract information from the unlabeled training set that has a close distribution to the test set. Since the resolution on the long axis varies significantly, we normalize the inter-slice distance along the long axis during training and testing, but rescale the final output to the original size for a fair comparison. A global DSC (combining all test cases into a single volume) is computed.
Directly applying the pre-trained model for segmentation reports an average accuracy of on the test set. After four generations, ATSO reports an average accuracy of (an over gain over the direct transfer baseline) which is higher than that of STSO () and the self-learning baseline () after the same number of generations. Note that is even higher than the accuracy () obtained from using labeled MSD data for training. Similar to the NIH experiments, we also obtain more accurate pseudo labels. After four generations, ATSO achieves a accuracy on the reference set, but both the self-learning baseline and STSO reports around . This aligns with our expectation: ATSO has the ability of transferring knowledge from a labeled dataset to another unlabeled, even when the data distributions differ considerably. Detailed results of transfer segmentation experiments from NIH to MSD are provided in Appendix B.
5.2 Transfer Segmentation across Natural Image Datasets
The second transfer learning scenario is defined between two natural image segmentation datasets. We use two popular datasets for autonomous driving, i.e., training the model on Cityscapes [4] and transferring it to Mapillary (https://www.mapillary.com/)111As a side comment, we also investigate the setting of semi-supervised learning within the Cityscapes dataset, and provide the results in Appendix C. The conclusions are similar to that reported in the NIH experiments.. The labeled training set contains finely annotated images from Cityscapes and the reference and test sets are from Mapillary, which have and images, respectively. Note that Cityscapes has semantic classes while Mapillary has . We train a model to infer classes and evaluate it on the reduced ground-truth on Mapillary by map each of the classes to one of the classes. The direct transfer method reports a mIOU of on the Mapillary test set that is dramatically lower than the number () on the Cityscapes test set, revealing a significant distribution gap between these two datasets.
We directly apply STSO and ATSO in this scenario, and obtain mIOU values of and (see Table 3), respectively. ATSO does not show advantages over STSO. This is mainly because -class segmentation is more challenging, so the noise in the pseudo labels may have been magnified to overwhelm the difference between STSO and ATSO. As an example, for the objects with a small ground-truth area (e.g., a train), the teacher model (not seeing the ground-truth) can easily ignore the entire object so that the pseudo label will guide the student models to make the same critical mistake. To alleviate this instability, we propose a solution that weakens the pseudo labels to get rid of the trouble. We further perform a pre-defined mapping to reduce the number of classes from to (i.e., rode, vehicle, person, vegetation, others) so as to reduce the risk of missing some class completely. In this reduced, -class segmentation task, ATSO reports a mIOU of which is significantly higher than the mIOU produced by STSO, .
We then use the -class pseudo label, generated by the final generation of either STSO or ATSO, to guide the target network. During the training procedure, the target network generates -class segmentation and then reduces it to classes for computing the loss function with respect to the pseudo labels. With only one training generation, the target model produces mIOU of and using the pseudo labels from STSO and ATSO, respectively, i.e., better pseudo labels lead to higher mIOU. Interestingly, when we continue fine-tuning the latter model (a mIOU) with -class pseudo labels, its performance is further boosted to , claiming an over gain beyond the direct transfer baseline (). Class-wise IOU numbers are detailed in Table 3. These results provide new insights to apply semi-supervised segmentation to the challenging datasets, in particular, with difficult semantic classes that can be easily missed. More details and results of the transfer experiments can be found in Appendix D.
Method | road | side | budg | wall | fence | pole | tr lt | tr sn | vegtr | terr | sky | pers | rider | car | truck | bus | train | motor | bike | mIoU |
transfer | 61.82 | 10.23 | 46.98 | 4.73 | 13.98 | 17.48 | 16.66 | 19.08 | 64.53 | 32.09 | 54.58 | 25.51 | 7.88 | 54.46 | 9.36 | 10.20 | 0.07 | 10.44 | 9.51 | 24.70 |
60.11 | 7.68 | 57.64 | 0.57 | 10.18 | 10.73 | 14.11 | 24.77 | 71.50 | 54.88 | 43.54 | 7.77 | 59.68 | 6.79 | 20.82 | 0.02 | 21.28 | 17.18 | 28.11 | ||
62.08 | 9.64 | 0.90 | 8.79 | 6.84 | 15.90 | 26.02 | 71.50 | 35.48 | 45.61 | 15.89 | 66.44 | 5.83 | 15.89 | 0.05 | 14.47 | 27.41 | 28.26 | |||
69.87 | 9.60 | 40.31 | 14.42 | 14.78 | 20.90 | 21.49 | 83.28 | 25.94 | 27.15 | 51.27 | 9.11 | 80.14 | 11.60 | 11.15 | 0.43 | 22.16 | 25.48 | |||
40.24 | 21.62 | 25.40 | 16.75 | 30.09 | 56.83 | 14.27 | 82.91 | 23.25 | 15.78 | 2.22 | 28.09 | 32.68 | 34.36 | |||||||
80.01 | 7.37 | 42.87 | 13.64 | 19.82 | 23.46 | 20.03 | 84.81 | 34.79 | 19.72 |
6 Conclusions
In this paper, we investigate the problem of semi-supervised learning, in particular, using teacher-student optimization, in the scenario of medical image segmentation. The core discovery is that the self-learning process can fall into a trap named lazy learning which downgrades the quality of prediction in the reference set. To alleviate this issue, we propose a simple yet effective pipeline named asynchronous teacher-student optimization (ATSO) which (i) switches off continual learning and (ii) avoids any unlabeled sample to be used in two consecutive fine-tuning rounds. Experiments on a few public datasets verify the effectiveness of our approach in both intra-dataset and inter-dataset semi-supervised learning tasks.
Our research sheds light on a new direction to improve semi-supervised learning, i.e., design a better schedule of feeding unlabeled data to the model. We expect to generalize this methodology to a wider range of learning tasks, including those on natural images. Also, the ATSO method needs further investigation, in particular, to solve the problem that unlabeled training data still suffer lower accuracy than the unseen testing data, as observed in all experiments.
Broader Impact
This paper presents a novel algorithm to improve the accuracy of semi-supervised medical image segmentation. We summarize the potential impact of our work in the following aspects.
-
•
To the research community. We reveal an interesting phenomenon named lazy learning, and propose an asynchronous optimization method to alleviate it. This provides a new insight to the semi-supervised learning community. In particular, conventional approaches mainly focused on introducing priors to make use of the unlabeled data, and we point out that such prior can sometimes lead us to a local optimum. Our solution can be integrated with many self-learning algorithms.
-
•
To training with limited labeled data. In the current era, a large amount of data of different types and modalities pour in everyday. However, few of them have been annotated. It has become increasingly important for the AI algorithms to learn from unlabeled data, and our algorithm advances the research in this field. Intriguingly, the ATSO method can be applied to different kinds of image data, including medical scans and autonomous driving data.
-
•
To the downstream engineers. With two examples (medical image and street image segmentation) studied in the paper and our code released, it will become easier for engineers to deploy our algorithm to other scenarios. While this may help to develop AI-based applications, there exist risks that some engineers, with relatively less knowledge in deep learning, can deliberately use the algorithm, e.g., without considering the data distribution of the reference set, which may actually harm the performance of the designed system.
-
•
To the society. There is a long-lasting debate on the impact that AI can bring to the human society. In particular, applying AI algorithms to medical care has become increasingly popular recently. Our algorithm can assist constructing a better system for medical image analysis, which is important to the scenarios that human expertise is limited or expensive, or in a public health event like the pandemic of COVID-19. However, we also notice that our algorithm can help to use personal medical data even if most of them have not been annotated. This raises concerns on privacy. Therefore, in general, it can bring both beneficial and harmful impacts and it really depends on the motivation of the users.
We also encourage the community to investigate the following problems.
-
1.
Is there any better solution for asynchronous optimization? The current method suffers the problem that being able to use only half of reference data in each generation.
-
2.
Are there any deeper insights of lazy learning? Is it related to some intrinsic difficulties of optimization that were not studied in this paper?
-
3.
Is it possible to generalize this algorithm to other vision tasks (e.g., image classification, object detection, etc.) or a generalized machine learning scenario?
References
- [1] A. Blum and T. Mitchell. Combining labeled and unlabeled data with co-training. In Annual Conference on Computational Learning Theory, 1998.
- [2] L. C. Chen, G. Papandreou, I. Kokkinos, Kevin Murphy, and A. L Yuille. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2018.
- [3] O. Cicek, A. Abdulkadir, S. S Lienkamp, T. Brox, and O. Ronneberger. 3d u-net: learning dense volumetric segmentation from sparse annotation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 2016.
- [4] M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and B. Schiele. The cityscapes dataset for semantic urban scene understanding. In 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 3213–3223, 2016.
- [5] T. Furlanello, Z. Lipton, M. Tschannen, L. Itti, and A. Anandkumar. Born-again neural networks. In International Conference on Machine Learning, 2018.
- [6] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition, 2016.
- [7] G. E Hinton, O. Vinyals, and J. Dean. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531, 2015.
- [8] Fabian Isensee, Paul F Jger, Simon A. A Kohl, Jens Petersen, and Klaus H Maier-Hein. Automated design of deep learning methods for biomedical image segmentation. 2019.
- [9] A. Krizhevsky, I. Sutskever, and G. E Hinton. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, pages 1097–1105, 2012.
- [10] Y. LeCun, Y. Bengio, and G. E. Hinton. Deep learning. Nature, 2015.
- [11] X. Li, L. Yu, H. Chen, C. W. Fu, and P. A. Heng. Semi-supervised skin lesion segmentation via transformation consistent self-ensembling model. In British Machine Vision Conference, 2018.
- [12] S. Liu, D. Xu, K. S Zhou, O. Pauly, S. Grbic, T. Mertelmeier, J. Wicklein, A. Jerebko, W. Cai, and D. Comaniciu. 3d anisotropic hybrid network: Transferring convolutional features from 2d images to 3d anisotropic volumes. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 2018.
- [13] J. Long, E. Shelhamer, and T. Darrell. Fully convolutional networks for semantic segmentation. In Computer Vision and Pattern Recognition, 2015.
- [14] Y. Man, Y. Huang, J. Feng, X. Li, and F. Wu. Deep q learning driven ct pancreas segmentation with geometry-aware u-net. IEEE Transactions on Medical Imaging, 2019.
- [15] F. Milletari, N. Navab, and S. A. Ahmadi. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In International Conference on 3D Vision, 2016.
- [16] T. Ni, L. Xie, H. Zheng, E. K Fishman, and A. L Yuille. Elastic boundary projection for 3d medical image segmentation. In Computer Vision and Pattern Recognition, 2019.
- [17] Mathias Perslev, Erik Bjrnager Dam, Akshay Pai, and Christian Igel. One network to segment them all: A general, lightweight system for accurate 3d medical image segmentation. 2019.
- [18] S. Qiao, W. Shen, Z. Zhang, B. Wang, and A. L Yuille. Deep co-training for semi-supervised image recognition. In European Conference on Computer Vision, 2018.
- [19] O. Ronneberger, P. Fischer, and T. Brox. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015.
- [20] C. Rosenberg, M. Hebert, and H. Schneiderman. Semi-supervised self-training of object detection models. In Winter Conference on Applications of Computer Vision, 2005.
- [21] H. R Roth, L. Lu, A. Farag, H. C. Shin, J. Liu, E. B Turkbey, and R. M Summers. Deeporgan: Multi-level deep convolutional networks for automated pancreas segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015.
- [22] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. International Conference on Learning Representations, 2015.
- [23] R. T. Sousa and J. Gama. Comparison between co-training and self-training for single-target regression in data streams using amrules. 2017.
- [24] S. Sun. A survey of multi-view machine learning. Neural Computing and Applications, 23(7-8):2031–2038, 2013.
- [25] N. Tajbakhsh, J. Y Shin, S. R Gurudu, Todd R Hurst, C. B Kendall, M. B Gotway, and J. Liang. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Transactions on Medical Imaging, 2016.
- [26] A. Tarvainen and H. Valpola. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Advances in neural information processing systems, 2017.
- [27] X. Wang, Y. Peng, L. Lu, Z. Lu, M. Bagheri, and R. M Summers. Chestx-ray8: Hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. In Computer Vision and Pattern Recognition, 2017.
- [28] Y. Xia, F. Liu, D. Yang, J. Cai, L. Yu, Z. Zhu, D. Xu, A. L Yuille, and H. Roth. 3d semi-supervised learning with uncertainty-aware multi-view co-training. In Winter Conference on Applications of Computer Vision, 2020.
- [29] Y. Xia, L. Xie, F. Liu, Z. Zhu, E. K Fishman, and A. L Yuille. Bridging the gap between 2d and 3d organ segmentation with volumetric fusion net. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 2018.
- [30] C. Xu, D. Tao, and C. Xu. A survey on multi-view learning. arXiv preprint arXiv:1304.5634, 2013.
- [31] C. Yang, L. Xie, S. Qiao, and A. L Yuille. Training deep neural networks in generations: A more tolerant teacher educates better students. In AAAI Conference on Artificial Intelligence, 2019.
- [32] C. Yang, L. Xie, C. Su, and A. L Yuille. Snapshot distillation: Teacher-student optimization in one generation. In Computer Vision and Pattern Recognition, 2019.
- [33] Q. Yu, L. Xie, Y. Wang, Y. Zhou, E. K Fishman, and A. L Yuille. Recurrent saliency transformation network: Incorporating multi-stage visual cues for small organ segmentation. In Computer Vision and Pattern Recognition, 2018.
- [34] Y. Zhang, T. Xiang, T. M Hospedales, and H. Lu. Deep mutual learning. In Computer Vision and Pattern Recognition, 2018.
- [35] Y. Zhou, Z. Li, S. Bai, C. Wang, X. Chen, M. Han, E. K Fishman, and A. L Yuille. Prior-aware neural network for partially-supervised multi-organ segmentation. In International Conference on Computer Vision, 2019.
- [36] Y. Zhou, Y. Wang, P. Tang, S. Bai, W. Shen, E. K Fishman, and A. L Yuille. Semi-supervised 3d abdominal multi-organ segmentation via deep multi-planar co-training. In Winter Conference on Applications of Computer Vision, 2019.
- [37] Y. Zhou, L. Xie, W. Shen, Y. Wang, E. K Fishman, and A. L Yuille. A fixed-point model for pancreas segmentation in abdominal ct scans. In International Conference on Medical Image Computing and Computer-Assisted Intervention, 2017.
- [38] X. J Zhu. Semi-supervised learning literature survey. Technical report, University of Wisconsin-Madison Department of Computer Sciences, 2005.
- [39] Z. Zhu, Y. Xia, W. Shen, E. K Fishman, and A. L Yuille. A 3d coarse-to-fine framework for automatic pancreas segmentation. In International Conference on 3D Vision, 2018.
Appendix A Additional Results of Semi-Supervised Learning on NIH
This part corresponds to Section 4.2 of the main article. The detailed results regarding the segmentation accuracy on the two reference subsets are summarized in Table 4. In each generation, we evaluate the two models on both and . We can see that in the first two generations, the model that gets trained on the other reference subset performs better, which verifies the lazy learning phenomenon. As the training process continues, this phenomenon becomes less significant, and this may be the main reason for the shrunk advantage of ATSO over STSO after the first two stages. In addition, this motivates us to consider if there is a better way to make use of data for the ATSO algorithm.
Generation | @ | @ | @ | @ | updated | @ |
---|---|---|---|---|---|---|
G0 | ||||||
G1 | ||||||
G2 | ||||||
G3 | ||||||
G4 | ||||||
G5 |
Generation | @ | @ | @ | @ | updated | @ |
---|---|---|---|---|---|---|
G0 | ||||||
G1 | ||||||
G2 | ||||||
G3 | ||||||
G4 |
Appendix B Additional Results of Transfer Learning from NIH to MSD
This part corresponds to Section 5.1 of the main article. The detailed results regarding the segmentation accuracy on the two reference subsets are summarized in Table 5. Conclusions are very similar to that observed in the semi-supervised experiments on NIH.
We also compare the segmentation results on the reference and test sets among the self-learning baseline, STSO and ATSO. Results are summarized in Table 6. We can observe that the performance of the self-learning baseline begins to degrade after two generations, while that of STSO and ATSO does not.
Generation | Self-learning | STSO | ATSO | |||
---|---|---|---|---|---|---|
@ | @ | @ | @ | @ | @ | |
G0 | ||||||
G1 | ||||||
G2 | ||||||
G3 | ||||||
G4 |
Appendix C Additional Results of Semi-Supervised Learning on Cityscapes
This part corresponds to Section 5.2 of the main article. We investigate the semi-supervised learning on the natural images for autonomous driving. We perform experiments on the Cityscapes dataset with finely-annotated images used for training and the remaining part of training data ( images, including finely-labeled and coarsely-labeled images) used for reference.
Results are summarized in Table 7. We still compare among four algorithms, namely, only training on the labeled data (transfer), the self-learning baseline, STSO, and ATSO. We can observe the same phenomenon as in the semi-supervised learning experiments in the NIH dataset, where ATSO improves consistently upon the other competitors. In particular, the overall improvement is over , and the improvement on hard classes is more significant.
Method | road | side | budg | wall | fence | pole | tr lt | tr sn | vegtr | terr | sky | pers | rider | car | truck | bus | train | motor | bike | mIoU |
transfer | 96.15 | 72.94 | 86.60 | 30.17 | 37.85 | 33.60 | 38.27 | 54.37 | 88.24 | 50.20 | 90.41 | 64.34 | 30.64 | 87.71 | 27.18 | 41.42 | 25.13 | 22.37 | 61.49 | 54.68 |
Self-learning | 97.60 | 81.12 | 90.53 | 45.02 | 51.58 | 60.56 | 73.12 | 91.38 | 55.67 | 94.35 | 77.70 | 53.51 | 92.56 | 47.60 | 59.24 | 48.30 | 55.37 | 72.64 | 68.48 | |
STSO | 51.64 | 53.45 | 58.46 | 92.55 | 45.30 | 59.68 | 43.45 | 53.17 | 68.72 | |||||||||||
ATSO | 97.64 | 81.74 | 89.97 | 43.18 | 51.09 | 54.74 | 69.72 | 91.15 | 94.03 | 76.18 | 51.90 | 70.99 |

Appendix D Additional Results of Transfer Learning from Cityscapes to Mapillary
This part corresponds to Section 5.2 of the main article. First, we report the self-learning baseline for the transfer learning experiments from Cityscapes to Mapillary, which reports an accuracy of (averaged over all classes). This is done without using the pre-defined mapping to reduce the number of classes from to . Note that this number is lower than the corresponding numbers of STSO () and ATSO ().
To show how our algorithm improves segmentation, we provide some typical examples in the transfer learning task from Cityscapes to Mapillary. We find that the direct transfer results are often below satisfaction, but the semi-supervised learning approaches can improve domain transfer performance dramatically. In particular, ATSO works best among all the solutions. In the first and third example, we find that ATSO5→19 is more stable at producing good results for the large objects. This is mainly because the learning process is stabilized by the pseudo labels generated on the reduced classes. On the other hand, when we tune the model by generating the -class pseudo label, the performance gain is often more significant on small objects. So, the proposed flowchart (starting from classes and then tuning on classes) is verified effective.