This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Colour augmentation for improved semi-supervised semantic segmentation

Geoff French1 [Uncaptioned image] and Michal Mackiewicz1 [Uncaptioned image]
1School of Computing Sciences, University of East Anglia, Norwich, UK
{g.french, m.mackiewicz}@uea.ac.uk
[Uncaptioned image] https://orcid.org/0000-0003-2868-2237[Uncaptioned image] https://orcid.org/0000-0002-8777-8880
Abstract

Consistency regularization describes a class of approaches that have yielded state-of-the-art results for semi-supervised classification. While semi-supervised semantic segmentation proved to be more challenging, a number of successful approaches have been recently proposed. Recent work explored the challenges involved in using consistency regularization for segmentation problems. In their self-supervised work Chen et al. found that colour augmentation prevents a classification network from using image colour statistics as a short-cut for self-supervised learning via instance discrimination. Drawing inspiration from this we find that a similar problem impedes semi-supervised semantic segmentation and offer colour augmentation as a solution, improving semi-supervised semantic segmentation performance on challenging photographic imagery.

1 INTRODUCTION

State-of-the-art computer vision results obtained using deep neural networks over the decade [Krizhevsky et al., 2012, He et al., 2016] rely on the availability of large training sets that consist of images and corresponding annotations. The annotation bottleneck resulting from the manual effort involved in producing these annotations can be partially mitigated by the application of semi-supervised learning. In contrast to traditional supervised learning in which all training samples have corresponding ground truth annotations, a semi-supervised learning algorithm is able to make used of un-annotated – or unsupervised – samples as well. The presents the possibility of using a dataset in which only a subset of the training samples have corresponding annotations.

Semantic segmentation is the task of identifying the type of object or material under each pixel in an image, assigning a class to every pixel. While efficient annotation tools [Maninis et al., 2018] can help, the cost of producing pixel-wise ground truth annotation can be significant, making the annotation bottleneck a particularly pressing issue for segmentation problems. The progress of semi-supervised semantic segmentation has lagged behind that of semi-supervised classification. [French et al., 2020] offer the challenging data distribution of semantic segmentation problems as an explanation.

The term consistency regularization [Oliver et al., 2018] refers to a class of approaches that have yielded state-of-the-art results for semi-supervised classification [Laine and Aila, 2017, Tarvainen and Valpola, 2017, Xie et al., 2019, Sohn et al., 2020] over the last few years. [French et al., 2020] explore the application of consistency regularization to semantic segmentation problems, developing a successful approach based on Cutmix [Yun et al., 2019]. They also find that plain geometric augmentation schemes used in prior semi-supervised classification approaches [Laine and Aila, 2017, Tarvainen and Valpola, 2017] frequently fail when applied to segmenting photographic imagery.

Recent work in self-supervised learning via instance discrimination trains a network for feature extraction using no ground truth labels at all. As with consistency regularization the network is encouraged to yield similar predictions – albeit image embeddings instead of probability vectors – given stochastically augmented variants of an unlabelled image. [Chen et al., 2020a] conducted a rigorous ablation study, finding that colour augmentation is essential to good performance. Without out, the network in effect cheats by using colour statistics as a short-cut for the image instance discrimination task used to train the network. Inspired by this, we find that a similar problem can hinder semi-supervised semantic segmentation. Our experiments demonstrate the problem by showing that it is alleviated by the use of colour augmentation.

Other recent approaches – namely Classmix [Olsson et al., 2021], DMT [Feng et al., 2021] and ReCo [Liu et al., 2021] – have significantly improved on the Cutmix based results of [French et al., 2020]. Our work builds on the Cutmix approach, demonstrating the effectiveness of colour augmentation. It is not our intent to present results competitive with Classmix and DMT, thus we acknowledge that our results are not state of the art.

2 BACKGROUND

2.1 Semi-supervised classification

The key idea behind consistency regularization based semi-supervised classification is clearly illustrated in the π\pi-model of Laine et al.  [Laine and Aila, 2017], in which a network is trained by minimizing both supervised and unsupervised loss terms. The supervised loss term applies traditional cross-entropy loss to supervised samples with ground truth annotations. Unsupervised samples are stochastically augmented twice and the unsupervised loss term encourages the network to predict consistent labels under augmentation.

The Mean Teacher model of Tarvainen et al.  [Tarvainen and Valpola, 2017] uses two networks; a teacher and a student. The weights of the teacher are an exponential moving average of those of the student. The student is trained using gradient descent as normal. The teacher network is used to generate pseudo-targets for unsupervised samples that the student is trained to match under stochastic augmentation.

The UDA approach of [Xie et al., 2019] adopted RandAugment [Cubuk et al., 2020]; a rich image augmentation scheme that chooses 2 or 3 image operations to apply from a menu of 14. While only one network is used instead of two, we note an important similarity with Mean Teacher; just as the teacher network is used to predict a pseudo target, UDA predicts a pseudo-target for an un-augmented image that is used as a training target for the same iamge with RandAugment applied.

The FixMatch approach of [Sohn et al., 2020] refines this approach further. They separate their augmentation scheme into weak – consisting of simple translations and horizontal flips – and strong that uses RandAugment. They predict hard pseudo-labels for weakly augmented unsupervised samples that are used as training targets for strongly augmented variants of the same samples.

2.2 Semi-supervised semantic segmentation

[Hung et al., 2018] and [Mittal et al., 2019] adopt GAN-based adversarial learning, using a discriminator network that distinguishes real from predicted segmentation maps to guide learning.

[Perone and Cohen-Adad, 2018] and [Li et al., 2018] two early applications of consistency regularisation to semantic segmentation that we are aware of. Both come from the medical imaging community, tackling MRI volume segmentation and skin lesion segmentation respectively. Both approaches use standard augmentation to provide perturbation, as in the π\pi-model [Laine and Aila, 2017] and Mean Teacher [Tarvainen and Valpola, 2017]. [Ji et al., 2019] developed a semi-supervised over-clustering approach that can be applied to natural photographic images, where the list of ground truth classes is highly constrained.

[French et al., 2020] analysed the problem of semantic segmentation, finding that it has a challenging data distribution to which the cluster assumption does not apply. They offer this as an explanation as to why consistency regularization had not been successfully applied to semantic segmentation of photographic images. They present an approach that drives the Mean Teacher [Tarvainen and Valpola, 2017] algorithm using an augmentation scheme based on Cutmix [Yun et al., 2019], achieving state of the art results.

Refer to caption
Figure 1: Illustration of Mean Teacher unsupervised consistency loss driven by standard augmentation for semantic segmentation problems. The path for a pixel on the neck of the cat leading from the input image xx is traced by yellows to the consistency loss map LconsL_{cons} (illustrated prior to computing the mean of the square), with the location of the pixel in each image identified by coloured crosses.

2.3 Self-supervised and unsupervised learning

Approaches based on contrastive learning [Henaff, 2020, He et al., 2020, Chen et al., 2020b, Chen et al., 2020a] train a residual network [He et al., 2016] using only unlabelled input images. Afterwards the network backbone (consisting of convolutional layers) is frozen and a linear model is trained in a supervised fashion using it’s feature representations as inputs and ground truth labels as targets. The resulting image classifiers – in which only the last linear layer was trained using ground truth labels – are able to achieve ImageNet results that are competitive with those obtained by traditional supervised learning in which the whole network is trained.

In contrast to prior work [Henaff, 2020] the MoCo model of He et al.  [He et al., 2020] simplified contrastive learning using standard augmentation to generate stochastically augmented variants of batches of unlabelled images. The network is encouraged to predict embeddings that are more similar for augmented variants of the same input image than for different images. The augmentation scheme used is very similar to the standard scheme used to train residual networks [He et al., 2016] and by Mean Teacher [Tarvainen and Valpola, 2017] for their ImageNet results. Chen et al.  [Chen et al., 2020a] conducted a rigorous ablation study of the augmentations used for contrastive learning, assessing the effectiveness of each augmentation operation. They found that strong colour augmentation is essential for good performance, as without it the network is able cheat by using image colour statisticts as a short-cut to discriminate between images, rather than having to focus on image content. Strong colour augmentation masks this signal, forcing the network to focus on the image content, extracting features suitable for accurate image classification and other downstream tasks.

We note the similarities between recent contrastive learning approaches and the Information Invariant Clustering approach of Ji et al.  [Ji et al., 2019], who also encourages consistency under stochastic augmentation.

The recent work of [Liu et al., 2021] adapt contrastive learning – typically used for classification – for semantic segmentation, achieving impressive results with very few labelled imgaes.

3 APPROACH

We adopt the approach and the codebase of the semi-supervised semantic segmentation work of [French et al., 2020] that combines Mean Teacher [Tarvainen and Valpola, 2017] with Cutmix [Yun et al., 2019]. We add colour augmentation and evaluate its’ effect. We will now describe the approaches that underpin our work.

Semi-supervised consistency loss: standard supervised cross entropy loss is combined with an unsupervised consistency loss term LconsL_{cons} that encourages consistent predictions under augmentation. In a classification scenario it measures the squared difference between probability predictions from the student network fθf_{\theta} and the teacher network gϕg_{\phi} given stochastically augmented variants x^\hat{x} and x~\tilde{x} of a sample xx:

Lcons=fθ(x^)gϕ(x~)2L_{cons}=\mathinner{\!\left\lVert f_{\theta}(\hat{x})-g_{\phi}(\tilde{x})\right\rVert}^{2} (1)

Semi-supervised segmentation driven with standard augmentation: Applying standard geometric augmentation – e.g. affine transformation – in both supervised and semi-supervised classification scenarios is straight forward. A class-preserving transformation is drawn randomly and applied to input image use in both the supervised and unsupervised loss terms. The predicted class probability vectors are unaffected by the transformation.

Applying geometric augmentation in segmentation scenarios requires a little more care as the classes of the pixels a semantic segmentation map – either a ground truth segmentation map yy used for the supervised loss term or a segmentation map gϕ(x)g_{\phi}(x) predicted by a teacher network used for an unsupervised loss term – correspond to the same pixels in the input image whose content they classify. This input-to-target pixel correspondance must be maintained.

When training a network we must apply the augmentation tαt_{\alpha} with identical parameters α\alpha to both the input image and segmentation map. For our supervised loss term this means computing the loss given the networks’ predictions fθ(tα(x))f_{\theta}(t_{\alpha}(x)) given the augmented input image tα(x)t_{\alpha}(x) and the augmented ground truth tα(y)t_{\alpha}(y). Following [Perone and Cohen-Adad, 2018] this can be adapted for the unsupervised loss term in a semi-supervised scenario by applying the geometric transformation tαt_{\alpha} to the input image prior to passing it to the student network and to the predicted segmentation from the teacher network (also illustrated in Figure 1):

Lcons=fθ(tα(x))tα(gϕ(x))2L_{cons}=\mathinner{\!\left\lVert f_{\theta}(t_{\alpha}(x))-t_{\alpha}(g_{\phi}(x))\right\rVert}^{2} (2)

Adding Cutmix to the mix: following [French et al., 2020] we use Cutmix to mix two input images xax_{a} and xbx_{b} to form a mixed image xmx_{m} using a blending mask mm: xm=xa(1m)+xbmx_{m}=x_{a}\odot(1-m)+x_{b}\odot m. The same blending mask is used to mix the segmentation maps predicted by the teacher network: ym=gϕ(xa)(1m)+gϕ(xb)my^{\prime}_{m}=g_{\phi}(x_{a})\odot(1-m)+g_{\phi}(x_{b})\odot m and the consistency loss term encourages the student predictions resulting from the mixed image to match the mixed segmentation maps:

Lcons=fθ(xm)ym2L_{cons}=\mathinner{\!\left\lVert f_{\theta}(x_{m})-y^{\prime}_{m}\right\rVert}^{2} (3)

3.1 Colour augmentation for consistency loss

Our choice to apply colour augmentation to the unsupervised loss term in a semi-supervised semantic segmentation setting was inspired by the thorough ablation study performed by Chen et al.  [Chen et al., 2020a] in which they explore the effects of augmentation in the similar setting of self-supervised learning. As stated in Section 2.3, they found that colour augmentation prevents the network from learning to use image colour statistics as a short-cut for image instance discrimination. Colour augmentation modifies the colour statistics sufficiently to prevent the network from using them to match image instances with one another trivially, forcing the network to focus on image content.

While [French et al., 2020] offer the challenging data distribution present in semantic segmentation problems as an explanation as to why consistency regularization driven by standard augmentation had yielded few successes when applied to photographic image datasets such as Pascal VOC 2012 [Everingham et al., 2012], we offer colour statistics as an alternative explanation.

The consistency loss term in equation 2 offers the opportunity for the network to minimize LconsL_{cons} using colour statistics. This is further illustrated in Figure 1, in which the yellow arrows follow a single pixel from the input image xx through both the student and teacher sides of the consistency loss term. Given the care taken to maintain the input-to-target pixel correspondence as stated in Section 3, most input pixels in xx (geometric augmentation can move some parts of an image outside the bounds of the image, hence correspondence for some pixels will be missing) will have corresponding pixels in the same locations in the prediction maps from both the student side fθ(tα(x))f_{\theta}(t_{\alpha}(x)) and the teacher side tα(gϕ(x))t_{\alpha}(g_{\phi}(x)). Given that the consistency loss term LconsL_{cons} penalises the network for giving inconsistent class predictions for each pixel, a simple way to minimize LconsL_{cons} is to predict the class of a pixel in the output segmentation maps using only the corresponding pixel in the input image, ignoring surrounding context. Thus, the network effectively clusters the colour of individual input pixels, rather than using surrounding context to identify the type of object that the pixel lies within.

Following [Chen et al., 2020a] we propose using colour augmentation to prevent the network from utilizing this short-cut. We acknowledge that [Ji et al., 2019] applied colour augmentation in an unsupervised semantic segmentation setting. While their codebase uses the same colour augmentation approach as [He et al., 2020] and [Chen et al., 2020a] they describe it simply as ‘photometric augmentation’ in their paper, giving little hint that it is in fact key to the success of consistency regularization based techniques in this problem domain.

\sim1/30 1/8 1/4 All
# labelled (100) (372) (744) (2975)
Results from other recent work, ImageNet pre-trained DeepLab v2 network
Baseline 56.2% 60.2% 66.0%
Adversarial 57.1% 60.5% 66.2%
s4GAN 59.3% 61.9% 65.8%
DMT 54.80% 63.06% 68.16%
Classmix 54.07% 61.35% 63.63%
Results from [French et al., 2020] and our results, ImageNet pre-trained DeepLab v2 network
Baseline 44.41% ±\pm 1.11 55.25% ±\pm 0.66 60.57% ±\pm 1.13 67.53% ±\pm 0.35
Cutout 47.21% ±\pm 1.74 57.72% ±\pm 0.83 61.96% ±\pm 0.99 67.47% ±\pm 0.68
   + colour aug. (ours) 48.28% ±\pm 1.98 58.30% ±\pm 0.73 62.59% ±\pm 0.60 67.93% ±\pm 0.36
CutMix 51.20% ±\pm 2.29 60.34% ±\pm 1.24 63.87% ±\pm 0.71 67.68% ±\pm 0.37
   + colour aug. (ours) 51.98% ±\pm 2.77 61.08% ±\pm 0.71 64.61% ±\pm 0.57 68.11% ±\pm 0.55
Table 1: Performance (mIoU) on Cityscapes validation set, presented as mean ±\pm std-dev computed from 5 runs. Other work: the results for ’Adversarial’ [Hung et al., 2018] and ’s4GAN’ [Mittal et al., 2019] are taken from [Mittal et al., 2019]. The results for DMT [Feng et al., 2021] and Classmix [Olsson et al., 2021] are from their respective works. Bold results in blue colour indicate results from other works that beat our best results. Our best results are in bold.
1/100 1/50 1/20 1/8 All
# labelled (106) (212) (529) (1323) (10582)
Results from other work with ImageNet pretrained DeepLab v2
Baseline 48.3% 56.8% 62.0% 70.7%
Adversarial 49.2% 59.1% 64.3% 71.4%
s4GAN+MLMT 60.4% 62.9% 67.3% 73.2%
DMT 63.04% 67.15% 69.92% 72.70% 74.75%
Classmix 54.18% 66.15% 67.77% 72.00%
Results from [French et al., 2020] + ours, ImageNet pre-trained DeepLab v2 network
Baseline 33.09% 43.15% 52.05% 60.56% 72.59%
Std. aug. 32.40% 42.81% 53.37% 60.66% 72.24%
    + colour aug. (ours) 46.42% 49.97% 57.17% 65.88% 73.21%
VAT 38.81% 48.55% 58.50% 62.93% 72.18%
    + colour aug. (ours) 40.05% 49.52% 57.60% 63.05% 72.29%
ICT 35.82% 46.28% 53.17% 59.63% 71.50%
    + colour aug. (ours) 49.14% 57.52% 64.06% 66.68% 72.91%
Cutout 48.73% 58.26% 64.37% 66.79% 72.03%
    + colour aug. (ours) 52.43% 60.15% 65.78% 67.71% 73.20%
CutMix 53.79% 64.81% 66.48% 67.60% 72.54%
    + colour aug. (ours) 53.19% 65.19% 67.65% 69.08% 73.29%
[French et al., 2020] + ours, ImageNet pre-trained DeepLab v3+ network
Baseline 37.95% 48.35% 59.19% 66.58% 76.70%
CutMix 59.52% 67.05% 69.57% 72.45% 76.73%
    + colour aug. (ours) 60.02% 66.84% 71.62% 72.96% 77.67%
[French et al., 2020] + ours, ImageNet pre-trained DenseNet-161 based Dense U-net
Baseline 29.22% 39.92% 50.31% 60.65% 72.30%
CutMix 54.19% 63.81% 66.57% 66.78% 72.02%
    + colour aug. (ours) 53.04% 62.67% 63.91% 67.63% 74.16%
[French et al., 2020] + ours, ImageNet pre-trained ResNet-101 based PSPNet
Baseline 36.69% 46.96% 59.02% 66.67% 77.59%
CutMix 67.20% 68.80% 73.33% 74.11% 77.42%
    + colour aug. (ours) 66.83% 72.30% 74.64% 75.40% 78.67%
Table 2: Performance (mIoU) on augmented Pascal VOC validation set, using same splits as Mittal et al[Mittal et al., 2019]. Other work: the results for ’Adversarial’ [Hung et al., 2018] and ’s4GAN’ [Mittal et al., 2019] are taken from [Mittal et al., 2019]. The results for DMT [Feng et al., 2021] and Classmix [Olsson et al., 2021] are from their respective works. Bold results in blue colour indicate results from other works that beat our best results. Our best results are in bold.
Baseline Std. aug. VAT ICT Cutout CutMix Fully sup.
(50) (2000)
Results from [Li et al., 2018] with ImageNet pre-trained DenseUNet-161
72.85% 75.31% 79.60%
Our results: Same ImageNet pre-trained DenseUNet-161
67.64% 71.40% 69.09% 65.45% 68.76% 74.57% 78.61%
±\pm 1.83 ±\pm 2.34 ±\pm 1.38 ±\pm 3.50 ±\pm 4.30 ±\pm 1.03 ±\pm 0.36
    + colour augmentation
73.61% 61.94% 50.93% 73.70% 74.51%
±\pm 2.40 ±\pm 6.72 ±\pm 7.16 ±\pm 2.59 ±\pm 1.95
Table 3: Performance on ISIC 2017 skin lesion segmentation validation set, measured using the Jaccard index (IoU for lesion class). Presented as mean ±\pm std-dev computed from 5 runs. All baseline and semi-supervised results use 50 supervised samples. The fully supervised result (’Full’) uses all 2000.

4 EXPERIMENTS

Our experiments follow the same procedure as [French et al., 2020], using the same network architectures. We used the same hyper-parameters, with the exception of the consistency loss weight that we will discuss in Section 4.3.1.

The loss term L=Lsup+γLconsL=L_{sup}+\gamma L_{cons} that we minimize is a weighted sum of a supervised cross-entropy term LsupL_{sup} with the consistency loss LconsL_{cons}, with γ\gamma as the consistency loss weight.

4.1 Implementation

Our implementation is directly based on that of [French et al., 2020]. We add colour augmentation to their implementation of standard augmentation, ICT [Verma et al., 2019], VAT [Miyato et al., 2017], Cutout [DeVries and Taylor, 2017] and Cutmix based regularizers. This allows us to assess its effect on a variety of regularizers across three datasets; Cityscapes, Pascal VOC 2012 and the ISIC Skin Lesion segmentation dataset [Codella et al., 2018]. Our colour augmentation randomly modifies the brightness, contrast, saturation and hue of an input image and is implemented using the ColorJitter transformation from the torchvision [Chintala et al., 2017] package.

4.2 Cityscapes

Cityscapes is a photograpic image dataset of urban scenery captured from the perspective of a car. Its’ training set consists of 2975 images.

Our Cityscapes results are presented in Table 1 as mean intersection-over-union (mIoU) percentages, where higher is better. The addition of colour augmentation results in a slight improvement to the CutOut and CutMix results across the board.

4.3 Augmented Pascal VOC 2012

Pascal VOC 2012[Everingham et al., 2012] is a photographic image dataset consisting of various indoor and outdoor scenes. It consists of only 1464 training images, and thus we follow the lead of [Hung et al., 2018] and augment it using Semantic Boundaries[Hariharan et al., 2011], resulting in 10582 training images.

Our Pascal VOC 2012 experiments evaluate regularizers based on standard augmentation, ICT [Verma et al., 2019] and VAT [Miyato et al., 2017], Cutout and Cutmix as in [French et al., 2020].

Our results are presented in Table 2.

4.3.1 Consistency loss weight

We note that the effects of colour augmentation resulted in different optimal values for γ\gamma (consistency loss weight) than were used by [French et al., 2020]. When using standard geometric augmentation they found that a value of 0.003 was optimal, yielding a very slight improvement over the supervised baseline. Increasing γ\gamma caused performance to drop below that of the supervised baseline. We note that at 0.003, the consistency loss term would have little effect on training at all. When using colour augmentation, we were able to use a value of 1 for γ\gamma; the same as that used for the more successful Cutout and CutMix regularizers. This strongly suggests that without colour augmentation, a low value must be used for γ\gamma to suppress the effect of the pixel colour clustering short-cut hypothesized in Section 3.1.

We were also able to use a value of 1 – instead of 0.01 – for the ICT [Verma et al., 2019] based regularizer when using colour augmentation. For VAT we continue to use a weight of 0.1; we attribute this lower loss weight to the use of KL-divergence in VAT rather than mean squared error for the consistency loss.

Being able to use a single value for the consistency loss weight for all regularizers simplifies the use of our approach in practical applications.

4.4 ISIC 2017 skin lesion segmentation

The ISIC skin lesion segmentation dataset [Codella et al., 2018] consists of dermoscopy images focused on lesions set against skin. It has 2000 images in its training set and is a two-class (skin and lesion) segmentation problem, featuring far less variation than Cityscapes and Pascal. Our results are presented in Table 3.

While colour augmentation improved the performance of all regularizers on the Pascal dataset when using the DeepLab v2 architecture, the results for ISIC 2017 are less clear cut. It harms the performance of VAT and ICT, although we note that we increased the consistency loss weight of ICT to match the value used for Pascal. It yields a noticeable improvement when using standard augmentation and Cutout. Colour augmentation increases the variance of the accuracy when using CutMix, making it slightly less reliable. We hypothesized the the hue jittering component of the colour augmentation may harm performance in this benchmark as colour is a useful queue in lesion segmentation, so we tried disabling it when using ICT and VAT. This did not however improve colour augmentation results.

4.5 Comparison with other work

While we have demonstrated that colour augmentation can improve semi-supervised segmentation performance when using a simple consistency regularization based approach, we acknowledge that our results do not match those of the recent Classmix [Olsson et al., 2021], DMT [Feng et al., 2021] and ReCo [Liu et al., 2021] approaches that use more recent semi-supervised regularizers.

We also note that [Liu et al., 2021] focused on situations in which a very small number of labelled samples were used. As their work did not feature experiments with a comparable number of labelled samples to our own, we were unable to directly compare their results with ours in Tables 1 and  2.

5 DISCUSSION AND CONCLUSIONS

As observed by [French et al., 2020] prior work in the field of semi-supervised image classification attributed the success of consistency regularization based approaches to the smoothness assumption [Luo et al., 2018] or cluster assumption [Chapelle and Zien, 2005, Sajjadi et al., 2016, Shu et al., 2018, Verma et al., 2019]. Their analysis of the data distribution of semantic segmentation showed that the cluster assumption does not apply. Their successful application of an adapted CutMix regularizer to semi-supervised semantic segmentation demonstrated that the cluster assumption is in fact not a pre-requisite for successful semi-supervised learning. In view of this, they offered the explanation that the variety of augmentation used need to provide perturbations to samples that are sufficiently varied in order to constrain the orientation of the decision boundary in the absence of the low density regions required by the cluster assumption. CutMix succeeds due to offering more variety than standard geometric augmentation.

Our results indicate a more nuanced explanation. The positive results obtained from adding colour augmentation to standard geometric augmentation, combined with being able to use a consistent value of 1 for the consistency loss weight for all regularizers shows that it is in fact the pixel colour clustering short-cut that was hampering the effectiveness of standard geometric augmentation by itself, rather than a lack of variation. Our results showing CutMix comfortably out-performing standard geometric augmentation with colour augmentation does however show that CutMix adds variety that enables more effective learning.

The story presented by the ISIC 2017 results is less positive however. The augmentation used to drive the consistency loss term in a semi-supervised learning scenario must be class preserving. Modifying an unsupervised sample such that its class changes will cause the consistency loss term to encourage consistent predictions across the decision boundary, harming the performance of the classifier (see the toy 2D examples in [French et al., 2020] for a more thorough exploration of this). In light of this, practitioners should carefully consider whether colour augmentation could alter the ground truth class of a sample. We offer this as an explanation of the inconsistent effect of colour augmentation on the ISIC 2017 dataset in which the colour of lesions is an important signal.

ACKNOWLEDGEMENTS

This work was in part funded under the European Union Horizon 2020 SMARTFISH project, grant agreement no. 773521. Much of the computation required by this work was performed on the University of East Anglia HPC Cluster. We would like to thank Jimmy Cross, Amjad Sayed and Leo Earl.

REFERENCES

  • [Chapelle and Zien, 2005] Chapelle, O. and Zien, A. (2005). Semi-supervised classification by low density separation. In AISTATS, volume 2005, pages 57–64. Citeseer.
  • [Chen et al., 2020a] Chen, T., Kornblith, S., Norouzi, M., and Hinton, G. (2020a). A simple framework for contrastive learning of visual representations. In Proceedings of the 37th International Conference on Machine Learning, volume 119, pages 1597–1607. PMLR.
  • [Chen et al., 2020b] Chen, X., Fan, H., Girshick, R., and He, K. (2020b). Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297.
  • [Chintala et al., 2017] Chintala, S. et al. (2017). Pytorch.
  • [Codella et al., 2018] Codella, N. C., Gutman, D., Celebi, M. E., Helba, B., Marchetti, M. A., Dusza, S. W., Kalloo, A., Liopyris, K., Mishra, N., Kittler, H., et al. (2018). Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (isbi), hosted by the international skin imaging collaboration (isic). In 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), pages 168–172. IEEE.
  • [Cubuk et al., 2020] Cubuk, E. D., Zoph, B., Shlens, J., and Le, Q. (2020). Randaugment: Practical automated data augmentation with a reduced search space. In Advances in Neural Information Processing Systems, volume 33, pages 18613–18624.
  • [DeVries and Taylor, 2017] DeVries, T. and Taylor, G. W. (2017). Improved regularization of convolutional neural networks with cutout. CoRR, abs/1708.04552.
  • [Everingham et al., 2012] Everingham, M., Van Gool, L., Williams, C. K. I., Winn, J., and Zisserman, A. (2012). The PASCAL Visual Object Classes Challenge 2012 (VOC2012) Results. http://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html.
  • [Feng et al., 2021] Feng, Z., Zhou, Q., Gu, Q., Tan, X., Cheng, G., Lu, X., Shi, J., and Ma, L. (2021). Dmt: Dynamic mutual training for semi-supervised learning. CoRR, abs/2004.08514.
  • [French et al., 2020] French, G., Laine, S., Aila, T., Mackiewicz, M., and Finlayson, G. (2020). Semi-supervised semantic segmentation needs strong, varied perturbations. In Proceedings of the British Machine Vision Conference (BMVC). BMVA Press.
  • [Hariharan et al., 2011] Hariharan, B., Arbeláez, P., Bourdev, L., Maji, S., and Malik, J. (2011). Semantic contours from inverse detectors. In International Conference on Computer Vision, pages 991–998.
  • [He et al., 2020] He, K., Fan, H., Wu, Y., Xie, S., and Girshick, R. (2020). Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9729–9738.
  • [He et al., 2016] He, K., Zhang, X., Ren, S., and Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 770–778.
  • [Henaff, 2020] Henaff, O. (2020). Data-efficient image recognition with contrastive predictive coding. In International Conference on Machine Learning, pages 4182–4192. PMLR.
  • [Hung et al., 2018] Hung, W.-C., Tsai, Y.-H., Liou, Y.-T., Lin, Y.-Y., and Yang, M.-H. (2018). Adversarial learning for semi-supervised semantic segmentation. CoRR, abs/1802.07934.
  • [Ji et al., 2019] Ji, X., Henriques, J. F., and Vedaldi, A. (2019). Invariant information clustering for unsupervised image classification and segmentation. In Proceedings of the IEEE International Conference on Computer Vision, pages 9865–9874.
  • [Krizhevsky et al., 2012] Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems 25, pages 1097–1105.
  • [Laine and Aila, 2017] Laine, S. and Aila, T. (2017). Temporal ensembling for semi-supervised learning. In International Conference on Learning Representations.
  • [Li et al., 2018] Li, X., Yu, L., Chen, H., Fu, C.-W., and Heng, P.-A. (2018). Semi-supervised skin lesion segmentation via transformation consistent self-ensembling model. In British Machine Vision Conference.
  • [Liu et al., 2021] Liu, S., Zhi, S., Johns, E., and Davison, A. J. (2021). Bootstrapping semantic segmentation with regional contrast. arXiv preprint arXiv:2104.04465.
  • [Luo et al., 2018] Luo, Y., Zhu, J., Li, M., Ren, Y., and Zhang, B. (2018). Smooth neighbors on teacher graphs for semi-supervised learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 8896–8905.
  • [Maninis et al., 2018] Maninis, K.-K., Caelles, S., Pont-Tuset, J., and Van Gool, L. (2018). Deep extreme cut: From extreme points to object segmentation. In Computer Vision and Pattern Recognition (CVPR).
  • [Mittal et al., 2019] Mittal, S., Tatarchenko, M., and Brox, T. (2019). Semi-supervised semantic segmentation with high-and low-level consistency. IEEE Transactions on Pattern Analysis and Machine Intelligence.
  • [Miyato et al., 2017] Miyato, T., Maeda, S.-i., Koyama, M., and Ishii, S. (2017). Virtual adversarial training: a regularization method for supervised and semi-supervised learning. arXiv preprint arXiv:1704.03976.
  • [Oliver et al., 2018] Oliver, A., Odena, A., Raffel, C., Cubuk, E. D., and Goodfellow, I. J. (2018). Realistic evaluation of semi-supervised learning algorithms. In International Conference on Learning Representations.
  • [Olsson et al., 2021] Olsson, V., Tranheden, W., Pinto, J., and Svensson, L. (2021). Classmix: Segmentation-based data augmentation for semi-supervised learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 1369–1378.
  • [Perone and Cohen-Adad, 2018] Perone, C. S. and Cohen-Adad, J. (2018). Deep semi-supervised segmentation with weight-averaged consistency targets. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, pages 12–19. Springer.
  • [Sajjadi et al., 2016] Sajjadi, M., Javanmardi, M., and Tasdizen, T. (2016). Mutual exclusivity loss for semi-supervised deep learning. In 23rd IEEE International Conference on Image Processing, ICIP 2016. IEEE Computer Society.
  • [Shu et al., 2018] Shu, R., Bui, H., Narui, H., and Ermon, S. (2018). A DIRT-t approach to unsupervised domain adaptation. In International Conference on Learning Representations.
  • [Sohn et al., 2020] Sohn, K., Berthelot, D., Carlini, N., Zhang, Z., Zhang, H., Raffel, C. A., Cubuk, E. D., Kurakin, A., and Li, C.-L. (2020). Fixmatch: Simplifying semi-supervised learning with consistency and confidence. In Advances in Neural Information Processing Systems, volume 33, pages 596–608.
  • [Tarvainen and Valpola, 2017] Tarvainen, A. and Valpola, H. (2017). Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In Advances in neural information processing systems, pages 1195–1204.
  • [Verma et al., 2019] Verma, V., Lamb, A., Kannala, J., Bengio, Y., and Lopez-Paz, D. (2019). Interpolation consistency training for semi-supervised learning. CoRR, abs/1903.03825.
  • [Xie et al., 2019] Xie, Q., Dai, Z., Hovy, E., Luong, M.-T., and Le, Q. V. (2019). Unsupervised data augmentation. arXiv preprint arXiv:1904.12848.
  • [Yun et al., 2019] Yun, S., Han, D., Oh, S. J., Chun, S., Choe, J., and Yoo, Y. (2019). Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE International Conference on Computer Vision, pages 6023–6032.