Understanding and Improving Robustness of Vision Transformers through Patch-based Negative Augmentation
Abstract
We investigate the robustness of vision transformers (ViTs) through the lens of their special patch-based architectural structure, i.e., they process an image as a sequence of image patches. We find that ViTs are surprisingly insensitive to patch-based transformations, even when the transformation largely destroys the original semantics and makes the image unrecognizable by humans. This indicates that ViTs heavily use features that survived such transformations but are generally not indicative of the semantic class to humans. Further investigations show that these features are useful but non-robust, as ViTs trained on them can achieve high in-distribution accuracy, but break down under distribution shifts. From this understanding, we ask: can training the model to rely less on these features improve ViT robustness and out-of-distribution performance? We use the images transformed with our patch-based operations as negatively augmented views and offer losses to regularize the training away from using non-robust features. This is a complementary view to existing research that mostly focuses on augmenting inputs with semantic-preserving transformations to enforce models’ invariance. We show that patch-based negative augmentation consistently improves robustness of ViTs on ImageNet based robustness benchmarks across 20+ different experimental settings. Furthermore, we find our patch-based negative augmentation are complementary to traditional (positive) data augmentation techniques and batch-based negative examples in contrastive learning.
1 Introduction
Building vision models that are robust, i.e., that maintain accuracy even on unexpected and out-of-distribution images, is increasingly a requirement to trusting vision models and a strong benchmark for progress in the field. Recently, Vision Transformers (ViTs, Dosovitskiy et al. (2021)) sparked great interest in the literature, as a radically new model architecture offering significant accuracy improvements and with hope of new robustness benefits. Over the past decade, there has been extensive work on understanding the robustness of convolution-based neural architectures, as the dominant design for visual tasks; researchers have explored adversarial robustness (Szegedy et al., 2013), domain generalization (Xiao et al., 2021; Khani & Liang, 2021), feature biases (Brendel & Bethge, 2019; Geirhos et al., 2018; Hermann et al., 2020). As a result, with the new promise of vision transformers, it is critical to understand their properties and in particular their robustness. Recent early studies (Naseer et al., 2021; Paul & Chen, 2021; Bhojanapalli et al., 2021) have found ViTs be more robust than ConvNets in some scenarios, with the hypothesis that the non-local attention based interactions enabled ViTs to capture more global and semantic features. In contrast, we add to this line of research showing a different side of the challenge: we find that ViTs are still vulnerable to relying on non-robust features, and propose an algorithm to mitigate it for better out-of-distribution performance.
Specifically, we first conduct a study on the robustness properties of ViTs and show that ViTs still rely on specific non-robust features. We start with the architectural traits of ViTs – ViTs operate on non-overlapping image patches and allow long range interaction between patches even in lower layers. It is hypothesized in recent studies (Naseer et al., 2021; Paul & Chen, 2021; Bhojanapalli et al., 2021) that the non-local attention based interactions contribute to better robustness of ViTs than ConvNets. To study the ability of ViTs to integrate global semantics structures across patches, we design and apply patch-based image transformations, such as random patch rotation, shuffling, and background-infilling (Figure 1). Those transformations destroy the spatial relationship between patches and corrupted the global semantics, and the resultant images are often visually unrecognizable. However, we find that ViTs are surprisingly insensitive to these transformations and can make highly accurate predictions on these transformed images. This suggests that ViTs use features that survive such transformations but are generally not indicative of the semantic class to humans. Going one step further, we find that those features are useful but not robust, as ViTs trained on them achieved high in-distribution accuracy, but suffered significantly on robustness benchmarks.
With this understanding of ViTs’ reliance on non-robust features captured by patch-based transformations, we aim to mitigate this shortcoming by answering the following questions: (a) how can we train ViTs to not rely on such features? and (b) will reducing reliance on such features meaningfully improve out-of-distribution performance and not sacrifice in-distribution accuracy? A majority of past robust training algorithms encourage the smoothness of model predictions on augmented images with semantic preserving transformations (Hendrycks et al., 2020; Cubuk et al., 2019). However, the patch-based transformations deliberately destroy the semantic meaning and only leave non-robust features. Taking inspiration from recent research on generative modeling (Sinha et al., 2020), we propose a family of robust training algorithms based on patch-based negative augmentations that regularize the training from relying on non-robust features surviving patch-based transformations. Through extensive evaluation on a wide set of ImageNet-based benchmarks, we find that our methods consistently improve the robustness of the trained ViTs. Furthermore, our patch-based negative augmentation can be combined with the traditional (positive) data augmentation to boost the performance further. With this we get a more complete picture: training models both to be insensitive to spurious changes (as in positive augmentation) but also to not rely on non-robust features (as in negative augmentation) together can meaningfully improve robustness of ViTs.

Our key contributions are as follows:
-
•
Understanding Non-Robust Features in ViT: We show that ViTs heavily rely on non-robust features surviving patch-based transformations but are not indicative of the semantic classes to humans.
-
•
Modeling: We propose a set of patch-based operations as negatively augmented views, complementary to existing works that focus on semantic-preserving (“positive”) augmentations, to regularize the training away from using these specific non-robust features;
-
•
Improved Robustness of ViT: We show across 20+ experimental settings that our proposed patch-based negative augmented views can consistently improve ViT’s robustness and complementary to “positive” augmentation techniques as well as batch-based negative examples in contrastive learning.
2 Preliminaries
Vision Transformers
Vision transformers (Dosovitskiy et al., 2021) are a family of architectures adapted from Transformers in natural language processing (Vaswani et al., 2017), which directly process visual tokens constructed from image patch embeddings. To construct visual tokens, ViT (Dosovitskiy et al., 2021) first splits an image into a grid of patches. Each patch is linearly projected into a hidden representation vector, and combined with a positional embedding. A learnable class token is also added. Transformers (Vaswani et al., 2017) are then directly applied on this set of visual and class token embeddings as if they are word embeddings in the original Transformer formulation. Finally, a linear projection of class token is used to calculate the class probability.
Model Variants
We consider ViT models pretrained on either ILSVRC-2012 ImageNet-1k, with 1.3 million images or ImageNet-21k, with 14 million images (Russakovsky et al., 2015). All models are fine-tuned on ImageNet-1k dataset. We adopt the notations used in (Dosovitskiy et al., 2021) to denote model size and input patch size. For example, ViT-B/16 denotes the “Base” model variant with input patch size .
Robustness Benchmarks
To evaluate models’ robustness, we mainly focus on three ImageNet-based robustness benchmarks, ImageNet-A (Hendrycks et al., 2019), ImageNet-C (Hendrycks & Dietterich, 2019a), ImageNet-R (Hendrycks et al., 2021). Specifically, ImageNet-A contains challenging natural images from a distribution unlike ImageNet training distribution, which can easily fool models to make a misclassification. ImageNet-C consists of 19 types of corruptions that are frequently encountered in natural images and each corruption has 5 levels of severity. It is widely used to measure models’ robustness under distributional shift. ImageNet-R is composed of images obtained by artistic rendition of ImageNet classes, e.g., cartoons, and is widely used to evaluate model’s robustness on out-of-distribution data.

3 Understanding the Robustness of Vision Transformers
Recent works (Naseer et al., 2021; Paul & Chen, 2021; Bhojanapalli et al., 2021) have shown that vision transformers achieve better robustness compared to standard convolutional networks. One explanation is that the attention mechanism can capture better global structures. To investigate if ViT has successfully taken advantage of the long range interactions between patches, we design a series of patch-based transformations which significantly destroys the global structure of images. The patch-based transformations (see Fig. 1) are:
-
•
Patch-based Shuffle (P-Shuffle): we randomly shuffle the input image patches to change their positions111P-Shuffle is equivalent to shuffling the position embeddings..
-
•
Patch-based Rotate (P-Rotate): we randomly select a rotation degree from the set and rotate each image patch independently.
-
•
Patch-based Infill (P-Infill): we replace the image patches in the center region of an image with the patches on the image boundary222For example, given an image with size , input patch size is and replace rate 0.25, we in total have 576 patches , where and denotes the row and column index and . The patches in the center are replaced by the remaining patches..
Each patch-based transformation is performed to a single image. We make sure the patch size of our patch-based transformation is a multiple of the input image patch of ViT so that the content within each patch is well-maintained. For P-infill, we use “replace rate” to denote the ratio of replaced patches in the center over the total number of patches in an image. Examples of transformed images are shown in Fig. 1 (see Appendix E for more examples). In most cases, it is challenging to recognize the semantic classes after those transformations.
Do ViTs rely on features not indicative of the semantic classes to humans? To validate if ViTs behave similarly as humans on these patch-based transformed images, we evaluate ViT models (Dosovitskiy et al., 2021) on these patch-based transformed image. Specifically, we apply each patch-based transformation to ImageNet-1k validation set and report the test accuracy of each ViT on the transformed images. The test accuracy is computed by using the semantic class of the corresponding original image as the ground-truth. As shown in Figure 2, the accuracies achieved by ViTs are significantly higher than random guessing (0.1). In addition, as shown in Figure 1, ViT gives these patch-based transformed images a highly confident prediction even when the transformation largely destroys the semantics and make the image unrecognizable by humans.333Similar patterns are observed when ViT models are pretrained on ImageNet-1k. This strongly indicates that ViT models heavily rely on the features that survive these transformations to make a prediction. However, these features are not indicative of the semantic class to humans.

Do features preserved in patch-based transformations impede robustness? Taking one step further, we want to know if the features preserved by simple patch-based transformations, which are not indicative of the semantic class to humans, result in robustness issues. To this end, we train a vision transformer, e.g., ViT-B/16, on patch-based transformed images with original semantic class assigned as their ground-truth. Note that all the training images are patch-based transformed images. In this way, we force the model to fully exploit the features preserved in patch-based transformations. In addition, we reuse all the training details and hyperparameters in (Dosovitskiy et al., 2021) to make sure the “only” difference between our models and the baseline ViT-B/16 (Dosovitskiy et al., 2021) is the training images. Then, we test the model on ImageNet-1k validation set and three robustness benchmarks, ImageNet-A, ImageNet-C and ImageNet-R without any transformation.
First, we can observe that for the baseline ViT-B/16, compared to in-distribution accuracy, all the out-of-distribution accuracies have suffered from a significant drop (the blue bars in Figure 3). This trend has been observed both for ViTs and convolution-based networks (Paul & Chen, 2021; Zhai et al., 2021). Second, if we compare the accuracy between the baseline model and models trained on patch-based transformations (i.e., the difference between the blue bar and one of the red/green/orange bars in Figure 3), we find that ViTs’ in-distribution accuracy drops only slightly, but the robustness drop is significant when models are trained on these patch-based transformations. Take P-Shuffle as an example, the model trained on patch-based shuffled images can still achieve 79.1 accuracy on ImageNet-1k, only 5 percentage point (pp) drop in in-distribution accuracy. In contrast, the accuracy drop on robustness datasets is much more significant, e.g., 17pp on ImageNet-R. The deterioration rate in robustness is close to 50 of the baseline ViT-B/16. This strongly suggests that the features preserved in patch-based transformations are sufficient for high-accurate in-distribution prediction but are not robust under distributional shifts.
Taking the above results together, we conclude that even though ViTs are shown to be more robust than ConvNets in previous studies (Naseer et al., 2021; Paul & Chen, 2021), they still heavily rely on features that are not indicative of the semantic classes to humans. These features, captured by patch-based transformations, are useful but non-robust, as ViTs trained on them achieve high in-distribution accuracy but suffer significantly on robustness benchmarks.
With this knowledge we now ask: based on these understandings, can we train ViTs to not rely on such non-robust features? And if we do, will it improve their robustness?
4 Improving the Robustness of Vision Transformers
Based on the key observations that the patch-based transformations encode features that contribute to the non-robustness of ViTs, we propose a negative augmentation procedure to regularize ViTs from relying on such features. To this end, we use the images transformed with our patch-based operations as negatively augmented views. Then we design negative loss regularizations to prevent the model from using those non-robust features preserved in patch-based transformations.
Specifically, given a clean image , we generate its negative view, denoted as , by applying a patch-based transformation to . We call it negative augmentation, in contrast with the standard (positive) augmentation that are semantic preserving. Let represent the cross-entropy loss function used to train a vision transformer with parameters , where is a minibatch of clean examples, and denotes the ground-truth label. The loss on negative views can be easily added to the cross-entropy loss via
(1) |
where is a coefficient balancing the importance between clean training data as well as patch-based negative augmentation. Below, we introduce uniform loss and loss to leverage negative views.
Uniform loss Many existing data augmentation techniques (Cubuk et al., 2019, 2020; Hendrycks et al., 2020) use one-hot labels for semantic-preserving augmented data to enforce the invariance of the model prediction. In contrast, the semantic classes of our generated patch-based negative augmented data are visually unrecognizable, as shown in Figure 1. Therefore, we propose to use uniform labels instead for those negative augmentations. Specifically, the loss function on negative views that we optimize at each training step can be formulated as:
(2) |
where denotes the uniform distribution: where is the total number of classes. denotes the function mapping the input image into the logit space.
Loss An alternative to pre-assuming labels for negative augmentation is to add the constraints on the logit space (or the space of predicted probability). Inspired by existing work (Kannan et al., 2018; Zhang et al., 2019; Hendrycks et al., 2020) which provides an extra regularization term encouraging similar logits between clean and “positive” augmented counterparts, we instead encourage the logits of clean examples and their corresponding negative augmentations to be far away. In this way, we prevent the model from relying on the non-robust features preserved in negative views. Specifically, we maximize the distance between the predicted probability of clean examples and their corresponding negative views. The loss on negative views, therefore, can be formulated as:
(3) |
Here the distance is computed over the predicted probability rather than the logits because empirically we observe that maximizing the difference of logits can cause numerical instability.
5 Experiments
Experimental setup We follow Dosovitskiy et al. (2021) to first pre-train all the models with image size and then fine-tune the models with a higher resolution . We reuse all their training hyper-parameters, including batch size, weight decay, and training epochs (see Appendix B for details). For the extra loss coefficient in Eqn. 1 we sweep it from the set and choose the model with the best hold-out validation performance. Please refer to Appendix C for the chosen hyperparameters for each model. We make sure our proposed models and our implemented baselines are trained with exactly the same settings for fair comparison.
5.1 Effective in improving robustness of vanilla ViTs
First, we apply our proposed patch-based transformations to a ViT model pre-trained and fine-tuned on ImageNet-1k. The extra loss regularization on negative views is used in both pre-training and fine-tuning stages to prevent the model from learning non-robust features preserved in patch-based transformations. We use “Transformation / Regularization” to denote a pair of patch-based negative augmentation and loss regularization. For examples, “P-Rotate / Uniform” means that we use P-Rotate to generate the negative views and use uniform loss to regularize the training. We display the results in Table 1, where we can clearly see that our proposed patch-based negative augmentation effectively improves the out-of-distribution robustness on ImageNet-C and ImageNet-R. Note the improvement is small over ImageNet-A when the baseline almost does not work, but it becomes more siginificant when the baseline improves (e.g., Table 5).
Larger improvement on small-scale datasets
We also apply our patch-based negative augmentation to ViT-B/16 on CIFAR-100 (Krizhevsky, 2009) in the fine-tuning stage and find that it can also significantly improve the robustness on the corrupted dataset CIFAR-100-C (Hendrycks & Dietterich, 2019b), which includes 19 different corruption types with 5 different corruption severities. For example, P-Shuffle / Uniform can boost the in-distribution accuracy of ViT-B/16 on CIFAR-100 from 91.8 to 92.6 (+0.8) and the robustness accuracy on CIFAR-100-C can also be improved from 74.6 to 77.0 (+2.4), as shown in Table 10 in Appendix.
Model | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
---|---|---|---|---|
ViT-B/32 (Dosovitskiy et al., 2021) | 72.5 | 3.7 | 46.7 | 19.4 |
+ P-Rotate / Uniform | 72.9 (+0.4) | 3.7 (+0.0) | 47.9 (+1.2) | 19.9 (+0.5) |
+ P-Rotate / L2 | 73.2 (+0.7) | 3.8 (+0.1) | 48.4 (+1.7) | 20.4 (+1.0) |
ViT-B/16 (Dosovitskiy et al., 2021) | 77.6 | 6.7 | 50.8 | 20.3 |
+ P-Rotate / Uniform | 78.2 (+0.6) | 7.0 (+0.3) | 52.4 (+1.6) | 21.4 (+1.1) |
+ P-Rotate / L2 | 77.8 (+0.2) | 6.7 (+0.0) | 51.6 (+0.8) | 21.0 (+0.7) |
5.2 Complementary to traditional (“positive”) data augmentation
To investigate if our proposed patch-based negative augmentation is complementary to traditional (“positive”) data augmentation, we apply our patch-based negative transformation on top of traditional data augmentation: Rand-Augment (Cubuk et al., 2020), which is widely used in vision transformers (Touvron et al., 2021; Mao et al., 2021), and AugMix (Hendrycks et al., 2020), which is specifically proposed to improve models’ robustness under distributional shift. Crucially, we follow (Hendrycks et al., 2020) to exclude transformations used in “positive” data augmentation which overlap with corruption types in ImageNet-C (Hendrycks & Dietterich, 2019a). Therefore, the set of transformations used in Rand-Augment and AugMix is disjoint with the corruptions in ImageNet-C.
When we combine “negative” and “positive” augmentation, the cross-entropy loss in Eqn. 1 is computed over “positive” examples using either Rand-Augment or AugMix. Meanwhile, the loss regularization on negative views in Eqn. 1 is computed over negatively transformed version of . That is: for , we apply our patch-based negative transformation to obtain its negative version and then use the negative example to compute the loss regularization . The positive data augmentation is only used in pre-training stage as we observe it is slightly better than using them for both stages (Please see more detailed discussion in Appendix D.4) and Steiner et al. (2021) have the similar observation. Instead, we apply our negative augmentation in both stages, as it is the best design choice as discussed in Appendix D.3.
As shown in Table 2, we see that when our patch-based negative augmentations are applied to either Rand-Augment or AugMix, we can consistently improve the robustness of vision transformers across all three robustness benchmarks. This is particularly noteworthy as both Rand-Agument and AugMix are already designed to significantly improve the robustness of vision models. Yet, we see that patch-based negative augmentation provides further robustness benefits. This suggests that robustness of vision models was not adequately addressed by “positive” data augmentation and that patch-based negative augmentation is complementary to these traditional approaches.
To have a clear picture how negative augmentation improves robustness on data with different degrees of corruptions, we also display the top-1 accuracy on ImageNet-C with five different levels of corruption severity in Table 3. The general trend emerges: as the corruption level goes higher, the benefit of negative examples becomes larger. This suggests our proposed negative examples are especially useful when data is highly corrupted.
Model | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
---|---|---|---|---|
Rand-Augment (Cubuk et al., 2020) | 79.1 | 7.2 | 55.2 | 23.0 |
+ P-Shuffle / Uniform | 79.3 | 7.7 (+0.5) | 56.2 (+1.0) | 23.4 (+0.4) |
+ P-Rotate / Uniform | 79.3 | 8.1 (+0.9) | 56.4 (+0.8) | 23.8 (+0.8) |
+ P-Infill / Uniform | 79.2 | 7.8 (+0.6) | 56.4 (+1.2) | 24.0 (+1.0) |
+ P-Shuffle / L2 | 78.9 | 7.5 (+0.3) | 55.6 (+0.4) | 22.6 (-0.4) |
+ P-Rotate / L2 | 79.1 | 7.9 (+0.7) | 56.7 (+1.5) | 23.8 (+0.8) |
+ P-Infill / L2 | 78.8 | 7.4 (+0.2) | 55.4 (+0.2) | 23.2 (+0.2) |
AugMix (Hendrycks et al., 2020) | 78.8 | 7.7 | 57.8 | 24.9 |
+ P-Shuffle / Uniform | 79.2 | 8.0 (+0.3) | 58.6 (+0.8) | 25.7 (+0.8) |
+ P-Rotate / Uniform | 79.1 | 8.2 (+0.5) | 58.5 (+0.7) | 25.7 (+0.8) |
+ P-Infill / Uniform | 79.3 | 8.3 (+0.6) | 58.4 (+0.6) | 25.7 (+0.8) |
+ P-Shuffle / L2 | 78.8 | 7.9 (+0.2) | 58.3 (+0.5) | 25.7 (+0.8) |
+ P-Rotate / L2 | 79.0 | 8.3 (+0.6) | 58.8 (+1.0) | 26.0 (+1.1) |
+ P-Infill / L2 | 79.0 | 7.9 (+0.2) | 58.5 (+0.7) | 25.6 (+0.7) |
Model | ImageNet-C Corruption Severity Level | ||||
---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | |
Rand-Augment (Cubuk et al., 2020) | 70.4 | 63.7 | 57.9 | 48.2 | 36.1 |
+ P-Rotate / Uniform | 71.1 (+0.7) | 64.6 (+0.9) | 59.0 (+1.1) | 50.0 (+1.8) | 37.6 (+1.5) |
+ P-Rotate / L2 | 71.1 (+0.7) | 64.8 (+1.1) | 59.5 (+1.6) | 50.1 (+1.9) | 37.8 (+1.7) |
AugMix (Hendrycks et al., 2020) | 71.4 | 65.2 | 60.5 | 51.9 | 40.2 |
+ P-Rotate / Uniform | 71.7 (+0.3) | 65.7 (+0.5) | 61.1 (+0.6) | 52.7 (+0.8) | 41.4 (+1.2) |
+ P-Rotate / L2 | 71.9 (+0.5) | 66.0 (+0.8) | 61.5 (+1.0) | 52.9 (+1.0) | 41.6 (+1.4) |
5.3 Complementary to batch-based negative examples in contrastive loss
In addition, we also investigate if our proposed patch-based negative augmentation can be incorporated into the traditional contrastive loss formulation and if they are complementary and provide additional value on top of the batch-based negative examples traditionally used (Oord et al., 2018; Chen et al., 2020a). We will first describe how to use contrastive loss as a regularization term and then introduce how to integrate patch-based negative augmentation into contrastive loss.
Following supervised contrastive loss proposed in (Khosla et al., 2020), for an example , we create a positive set with all the examples in the minibatch sharing the same class as . The anchor is excluded from its positive set . Next, all the examples in the minibatch with a different class as are used as negative examples. Let the candidate set , the contrastive loss can be expressed as
(4) |
where is the temperature and computes the cosine similarity between and , and denotes the representation learned by the penultimate layer of the classifier. We do not use a learnable projection head as in contrastive representation learning (Chen et al., 2020a, b) to avoid extra network parameters. The contrastive loss is used as the regularization term in Eqn 1. We denote this stronger baseline as “Contrastive*”.
To integrate our proposed patch-based negative examples into contrastive loss, we expand the negative set in contrastive loss, which now composed of two types of negative examples: 1) all the examples in the minibatch with a different class as , 2) the patch-based negatively transformed images . Therefore, for each anchor , we can in total have negative pairs, where is the batch size and is the cardinality of the positive set . Accordingly, the candidate set in Eqn 4 becomes . In Table 4, we can see that even if we add the patch-based negative augmentation on top of this stronger contrastive baseline, we can still consistently achieve extra improvement across robustness benchmarks. This shows our proposed patch-based negative augmentation is also complementary to batch-based negative examples in contrastive loss in improving models’ robustness. Similar trend holds on other settings, see Table 14 in Appendix. Comparing to the other two negative loss regularizations, uniform and loss, we suggest readers to use our proposed contrastive loss in practice since it incorporates the extra benefit of constraining the embeddings of positive pairs to be similar and consistently performs better.
Model | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
---|---|---|---|---|
ViT-B/16 + Contrastive* | 78.7 | 8.1 | 53.5 | 22.8 |
ViT-B/16 + P-Rotate / Contrastive | 78.9 | 8.6 (+0.5) | 54.1 (+0.6) | 23.6 (+0.8) |
Rand-Augment + Contrastive* | 79.7 | 8.9 | 57.6 | 24.7 |
Rand-Augment + P-Rotate / Contrastive | 79.9 | 9.4 (+0.5) | 58.4 (+0.8) | 25.4 (+0.7) |
5.4 Robustness improvements even under larger pre-training datasets
Model | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
---|---|---|---|---|
Rand-Augment (Cubuk et al., 2020) | 84.40.0 | 28.70.2 | 67.20.0 | 38.70.1 |
+ P-Shuffle / Uniform | 84.5 | 29.9 | 67.7 | 38.9 |
+ P-Shuffle / L2 | 84.50.0 | 29.70.3 | 68.00.0 | 39.60.2 |
We further investigate if our proposed method can scale up to larger datasets and continues to be necessary and valuable. To this end, we test if our proposed patch-based negative augmentation still helps robustness when models are pre-trained on ImageNet-21k (10x larger than ImageNet-1k), where the baseline performance is higher. Take P-Shuffle as an example, we display the results in Table 5 with negative augmentation in both pre-training and fine-tuning stages. We see that even when the pre-training dataset is significantly increased, our patch-based negative augmentation can still further improve the robustness of ViT. This demonstrates that our approach is valuable at scale and improves models’ robustness from an angle orthogonal to larger pre-training data.
Mean and standard deviation: Considering training models from scratch on ImageNet-1k or 21k is very costly, we take the strongest baseline as an example to independently run multiple times. As shown in Table 5, the small standard deviations suggests the improvement is statistically significant. In addition, we have presented 20+ different experimental settings in the paper where our proposed patch-based negative examples consistently improves robustness.
5.5 Patch-based negative augmentation reduces texture bias
Geirhos et al. (2018) observed that unlike humans, CNNs rely on more local information (e.g., texture) rather than more global information (e.g., shape) to make a classification. Since our patch-based transformations largely destroy the global structure (e.g., shape), we want to investigate if the non-robust features surviving patch-based transformation overlap with local texture biases. To this end, we evaluate ViT-B/16 trained on patch-based transformations on Conflict Stimuli benchmark (Geirhos et al., 2018), and we see that ViTs trained only on patch-based transformation have a 4.9pp to 31.1pp increase on texture bias (Figure 5 in Appendix). This suggests that the useful but non-robust features preserved in patch-based transformation are indeed overlapped with the local texture bias. In addition, using our patch-negative augmentation can also to some extent reduce models’ reliance on local texture bias, e.g., we decrease the texture accuracy from 71.7 to 66.5 for ViT-B/16 (Table 6 in Appendix).
6 Discussions
Does ViT become more robust w.r.t. transformed images?
We further evaluate ViTs trained with our robust training algorithms on the patch-based transformed images. We found all three losses on negative views can successfully reduce the prediction accuracy of ViTs to be close to random guess with the original semantic classes as the ground-truth. In other words, our robust training algorithms make ViTs behave similarly as humans on those patch-based transformed images.

Are high confidence predictions w.r.t. patch transformed images due to non-overlapping patch embeddings?
To answer this, we test original ViT from Dosovitskiy et al. (2021) on the patch-transformed images where the patch size of the patch-based transformations is not perfectly aligned with the input patch size , e.g., is not a multiple of . This can approximate the scenario when patch embeddings are extracted from overlapped patches. As shown in Table 11 in Appendix, when the patch size of the patch-based transformation ( = 24) and the input patch size ( = 16) of ViT are not perfectly aligned, the test accuracy on the patch-based shuffled images decreases from 84.1 to 32.9. This is lower than the case that the patch sizes are aligned, e.g., the test accuracy is 57.8 when , but still significantly higher than humans (a random guess is close to 0.1) as well as comparable CNN-based network . Therefore, we conjecture the high confidence w.r.t. transformed images is partially from non-overlapping patch embedding in ViTs.
When does patch-based negative augmentation help?
To answer this question, we follow (Dosovitskiy et al., 2021) to visualize the attention maps of models trained with/without our proposed negative data augmentation. In Figure 4, we can see that our negative augmentation can effectively 1) help models attend to the object to make a correct prediction whereas the standard ViT fails (top two rows), and 2) mitigate models’ reliance on unrobust biases, e.g., local bias, color bias, texture bias, etc. For example, the cartoon Axolotl (third row) is misclassified by the standard ViT as Goldfish due to color similarity and the Starfish (bottom row) is incorrectly recognized as Ocarina because of local resemblance. This is aligned with other signals in previous sections that standard ViT relies on unrobust biases to make a prediction, whereas our negative augmentation effectively encourages the model to rely less on these unrobust features. Interestingly, for the bottom two rows, the standard ViT has successfully attended to the object in both cases but still failed to make correct predictions, indicating that the attention on the object does not ensure a model is relying on semantically meaningful and robust features to make a prediction.
7 Related Work
Vision transformers (Dosovitskiy et al., 2021; Touvron et al., 2021) are a family of Transformer models (Vaswani et al., 2017) that directly process visual tokens constructed from image patch embedding. Unlike convolutional neural networks (LeCun et al., 1989; Krizhevsky et al., 2012; He et al., 2016) that assume locality and translation invariance in their architectures, vision transformers have no such assumptions and can exchange information globally. A few recent studies find pretrained vision transformers are at least as robust as the ResNet counterparts (Bhojanapalli et al., 2021), and possibly more robust (Naseer et al., 2021; Paul & Chen, 2021). Our work studies a specific aspect of robustness pertaining patch-based visual tokens in ViT, and show it may lead to a generalization gap. Different from (Naseer et al., 2021) which also shows ViTs are insensitive to patch operations such as shuffle and occlusion, we design different types of patch-based transformations and develop deeper understanding that the features preserved in patch-based transformations are non-robust. Further, we propose a mitigation strategy based on these patch-based transformations to increase robustness of vision transformers. Another orthogonal line of improving robustness of vision transformers is to develop a deeper understanding of the self-attention mechanisms in vision transformers (Gu et al., 2022) and further advances the neural network architectures (Zhou et al., 2022).
Data augmentation is widely used in computer vision models to improve model performance (Howard, 2013; Szegedy et al., 2015; Cubuk et al., 2020, 2019; Touvron et al., 2021). However, most of the existing data augmentations are “positive” in the sense they assume the class semantic being preserved after the transformation. In this work, we explore “negative” data augmentation operations based on patches, where we encourage the representations of transformed example to be different from the original ones. Most related to our work in this direction is the work of Sinha et al. (2020). Although the concept of negative augmentation was proposed in their work, they only apply it for generative and unsupervised modeling without consideration of robustness. In contrast, our work focuses on discriminative and supervised modeling, and demonstrate how such negative examples can reveal specific robustness issues and such augmentation approaches can offer robustness improvements under large-scale pretraining settings.
Our work is also related to contrastive learning (Wu et al., 2018; Hjelm et al., 2019; Oord et al., 2018; He et al., 2020; Tian et al., 2020). The increasing number of negative pairs has shown to be important for representation learning in self-supervised contrastive learning (Chen et al., 2020a), where different images serve as negative examples for each other, and supervised contrastive learning (Khosla et al., 2020), where images with different classes are used as negative examples. Unlike the traditional setting of representation learning, our proposed contrastive loss serves as a regularization term with patch-based negative augmentations as extra negative data points.
8 Conclusion
Through this research we have found concrete evidences of ViTs relying on non-robust features to make predictions and shown that this reliance is limiting out-of-distribution robustness. This opens multiple exciting new lines of research. First, our methodology for analyzing the robustness of ViTs provides a valuable recipe for future research. Through designing patch-wise, semantic-destroying transformations that ViTs are insensitive to, we identified which non-robust features models rely on. Second, through negative augmentations during training we reduced the ViTs’ reliance on such non-robust features, and improved the out-of-distribution performance of ViTs significantly, without harming in-distribution accuracy. This approach shows the potential for further improving the robustness of ViTs. While we have identified multiple such non-robust features in ViTs, we believe discovering and addressing more is a promising direction for continued progress toward robust vision transformers and vision models in general.
Acknowledgements
The authors would like to thank Alexey Dosovitskiy, Lucas Beyer and Neil Houlsby for helpful discussions on vision transformer systems. We also want to thank the reviewers for their useful comments.
References
- Bhojanapalli et al. (2021) Srinadh Bhojanapalli, Ayan Chakrabarti, Daniel Glasner, Daliang Li, Thomas Unterthiner, and Andreas Veit. Understanding robustness of transformers for image classification. arXiv preprint arXiv:2103.14586, 2021.
- Brendel & Bethge (2019) Wieland Brendel and Matthias Bethge. Approximating cnns with bag-of-local-features models works surprisingly well on imagenet. In The International Conference on Learning Representations, 2019.
- Chen et al. (2020a) Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning, pp. 1597–1607. PMLR, 2020a.
- Chen et al. (2020b) Ting Chen, Simon Kornblith, Kevin Swersky, Mohammad Norouzi, and Geoffrey Hinton. Big self-supervised models are strong semi-supervised learners. Advances in Neural Information Processing Systems, 2020b.
- Cubuk et al. (2019) Ekin D Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V Le. Autoaugment: Learning augmentation strategies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 113–123, 2019.
- Cubuk et al. (2020) Ekin D Cubuk, Barret Zoph, Jonathon Shlens, and Quoc V Le. Randaugment: Practical automated data augmentation with a reduced search space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 702–703, 2020.
- Dosovitskiy et al. (2021) Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. International Conference on Learning Representations, 2021.
- Geirhos et al. (2018) Robert Geirhos, Patricia Rubisch, Claudio Michaelis, Matthias Bethge, Felix A Wichmann, and Wieland Brendel. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. In International Conference on Learning Representations, 2018.
- Gu et al. (2022) Jindong Gu, Volker Tresp, and Yao Qin. Are vision transformers robust to patch perturbations? In European Conference on Computer Vision, 2022.
- He et al. (2016) Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778, 2016.
- He et al. (2020) Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729–9738, 2020.
- Hendrycks & Dietterich (2019a) Dan Hendrycks and Thomas Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In International Conference on Learning Representations, 2019a.
- Hendrycks & Dietterich (2019b) Dan Hendrycks and Thomas Dietterich. Benchmarking neural network robustness to common corruptions and perturbations. In Proceedings of the International Conference on Learning Representations, 2019b.
- Hendrycks et al. (2019) Dan Hendrycks, Kevin Zhao, Steven Basart, Jacob Steinhardt, and Dawn Song. Natural adversarial examples. arXiv preprint arXiv:1907.07174, 2019.
- Hendrycks et al. (2020) Dan Hendrycks, Norman Mu, Ekin Dogus Cubuk, Barret Zoph, Justin Gilmer, and Balaji Lakshminarayanan. AugMix: A Simple Data Processing Method to Improve Robustness and Uncertainty. In International Conference on Learning Representations, 2020.
- Hendrycks et al. (2021) Dan Hendrycks, Steven Basart, Norman Mu, Saurav Kadavath, Frank Wang, Evan Dorundo, Rahul Desai, Tyler Zhu, Samyak Parajuli, Mike Guo, Dawn Song, Jacob Steinhardt, and Justin Gilmer. The many faces of robustness: A critical analysis of out-of-distribution generalization. International Conference on Computer Vision, 2021.
- Hermann et al. (2020) Katherine L Hermann, Ting Chen, and Simon Kornblith. The origins and prevalence of texture bias in convolutional neural networks. In Advances In Neural Information Processing Systems, 2020.
- Hjelm et al. (2019) R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. International Conference on Learning Representations, 2019.
- Howard (2013) Andrew G Howard. Some improvements on deep convolutional neural network based image classification. arXiv preprint arXiv:1312.5402, 2013.
- Kannan et al. (2018) Harini Kannan, Alexey Kurakin, and Ian Goodfellow. Adversarial logit pairing. arXiv preprint arXiv:1803.06373, 2018.
- Khani & Liang (2021) Fereshte Khani and Percy Liang. Removing spurious features can hurt accuracy and affect groups disproportionately. In Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, pp. 196–205, 2021.
- Khosla et al. (2020) Prannay Khosla, Piotr Teterwak, Chen Wang, Aaron Sarna, Yonglong Tian, Phillip Isola, Aaron Maschinot, Ce Liu, and Dilip Krishnan. Supervised contrastive learning. Advances In Neural Information Processing Systems, 2020.
- Kingma & Ba (2015) Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. CoRR, abs/1412.6980, 2015.
- Krizhevsky (2009) Alex Krizhevsky. Learning multiple layers of features from tiny images. Technical report, University of Toronto, 2009.
- Krizhevsky et al. (2012) Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. Advances in Neural Information Processing Systems, 25:1097–1105, 2012.
- LeCun et al. (1989) Yann LeCun, Bernhard Boser, John S Denker, Donnie Henderson, Richard E Howard, Wayne Hubbard, and Lawrence D Jackel. Backpropagation applied to handwritten zip code recognition. Neural Computation, 1(4):541–551, 1989.
- Mao et al. (2021) Xiaofeng Mao, Gege Qi, Yuefeng Chen, Xiaodan Li, Shaokai Ye, Yuan He, and Hui Xue. Rethinking the design principles of robust vision transformer. arXiv preprint arXiv:2105.07926, 2021.
- Naseer et al. (2021) Muzammal Naseer, Kanchana Ranasinghe, Salman Khan, Munawar Hayat, Fahad Shahbaz Khan, and Ming-Hsuan Yang. Intriguing properties of vision transformers. arXiv preprint arXiv:2105.10497, 2021.
- Oord et al. (2018) Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
- Paul & Chen (2021) Sayak Paul and Pin-Yu Chen. Vision transformers are robust learners. arXiv preprint arXiv:2105.07581, 2021.
- Russakovsky et al. (2015) Olga Russakovsky, Jia Deng, Hao Su, Jonathan Krause, Sanjeev Satheesh, Sean Ma, Zhiheng Huang, Andrej Karpathy, Aditya Khosla, Michael Bernstein, Alexander C. Berg, and Li Fei-Fei. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision (IJCV), 115(3):211–252, 2015. doi: 10.1007/s11263-015-0816-y.
- Sinha et al. (2020) Abhishek Sinha, Kumar Ayush, Jiaming Song, Burak Uzkent, Hongxia Jin, and Stefano Ermon. Negative data augmentation. In International Conference on Learning Representations, 2020.
- Steiner et al. (2021) Andreas Steiner, Alexander Kolesnikov, Xiaohua Zhai, Ross Wightman, Jakob Uszkoreit, and Lucas Beyer. How to train your vit? data, augmentation, and regularization in vision transformers. arXiv preprint arXiv:2106.10270, 2021.
- Szegedy et al. (2013) Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199, 2013.
- Szegedy et al. (2015) Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9, 2015.
- Tian et al. (2020) Yonglong Tian, Dilip Krishnan, and Phillip Isola. Contrastive multiview coding. In European Conference on Computer Vision, pp. 776–794. Springer, 2020.
- Touvron et al. (2021) Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Hervé Jégou. Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning, pp. 10347–10357. PMLR, 2021.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in Neural Information Processing Systems, pp. 5998–6008, 2017.
- Wu et al. (2018) Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3733–3742, 2018.
- Xiao et al. (2021) Kai Xiao, Logan Engstrom, Andrew Ilyas, and Aleksander Madry. Noise or signal: The role of image backgrounds in object recognition. International Conference on Learning Representations, 2021.
- Zhai et al. (2021) Xiaohua Zhai, Alexander Kolesnikov, Neil Houlsby, and Lucas Beyer. Scaling vision transformers, 2021.
- Zhang et al. (2019) Hongyang Zhang, Yaodong Yu, Jiantao Jiao, Eric P Xing, Laurent El Ghaoui, and Michael I Jordan. Theoretically principled trade-off between robustness and accuracy. International Conference on Machine Learning, 2019.
- Zhou et al. (2022) Daquan Zhou, Zhiding Yu, Enze Xie, Chaowei Xiao, Animashree Anandkumar, Jiashi Feng, and Jose M Alvarez. Understanding the robustness in vision transformers. In International Conference on Machine Learning, pp. 27378–27394. PMLR, 2022.
Checklist
-
1.
For all authors…
-
(a)
Do the main claims made in the abstract and introduction accurately reflect the paper’s contributions and scope? [Yes]
-
(b)
Did you describe the limitations of your work? [Yes]
-
(c)
Did you discuss any potential negative societal impacts of your work? [N/A] Not Applicable
-
(d)
Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes]
-
(a)
-
2.
If you are including theoretical results…
-
(a)
Did you state the full set of assumptions of all theoretical results? [Yes]
-
(b)
Did you include complete proofs of all theoretical results? [Yes]
-
(a)
-
3.
If you ran experiments…
-
(a)
Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] See Supplementary Materials.
-
(b)
Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Appendix.
-
(c)
Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] See Section 5.4.
-
(d)
Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [No]
-
(a)
-
4.
If you are using existing assets (e.g., code, data, models) or curating/releasing new assets…
-
(a)
If your work uses existing assets, did you cite the creators? [Yes]
-
(b)
Did you mention the license of the assets? [Yes]
-
(c)
Did you include any new assets either in the supplemental material or as a URL? [No]
-
(d)
Did you discuss whether and how consent was obtained from people whose data you’re using/curating? [N/A]
-
(e)
Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A]
-
(a)
-
5.
If you used crowdsourcing or conducted research with human subjects…
-
(a)
Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A]
-
(b)
Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A]
-
(c)
Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]
-
(a)
Appendix
Appendix A Patch-based Negative Data Augmentation Reduces Texture Bias

Pre-train on ImageNet-1k | Pre-train on ImageNet-21k | ||
Model | Texture Accuracy | Model | Texture Accuracy |
ViT-B/16 | 71.7 | Rand-Augment | 57.5 |
+ P-Rotate / Uniform | 66.5 | + P-Shuffle / Uniform | 56.4 |
+ P-Rotate / L2 | 67.2 | + P-Shuffle / L2 | 54.7 |
Appendix B Training Details
We follow (Dosovitskiy et al., 2021) to train each model using Adam (Kingma & Ba, 2015) optimizer with = 0.9, = 0.999 for pre-training and SGD with momentum for fine-tuning. The batch size is set to be 4096 for pre-training and 512 for fine-tuning. All models are trained with 300 epochs on ImageNet-1k and 90 epochs on ImageNet-21k in the pre-training stage. In the fine-tuning stage, all models are trained with 20k steps except the models pretrained from ImageNet-1k without Rand-Augment (Cubuk et al., 2020) or Augmix (Hendrycks et al., 2020), which we train them with 8k steps. The learning rate warm-up is set to be 10k steps. Dropout is used for both pre-training and fine-tuning with dropout rate 0.1. If the training dataset is ImageNet-1k, we additionally apply gradient clipping at global norm .
Pre-train Dataset | Stage | Base LR | LR Decay | Weight Decay | Label Smoothing |
---|---|---|---|---|---|
ImageNet-1K | Pre-train | ‘cosine’ | None | ||
ImageNet-21k | Pre-train | ‘linear’ | 0.03 | ||
ImageNet-1K | Fine-tune | 0.01 | ‘cosine’ | None | None |
ImageNet-21K | Fine-tune | 0.03 | ‘cosine’ | None | None |
Model | Pre-train Dataset | Training stage | Hyperparameter |
---|---|---|---|
Rand-Augment + P-Shuffle / Uniform | ImageNet-1k | Pre-train | 1.0 |
Rand-Augment + P-Shuffle / Contrastive | ImageNet-1k | Pre-train | 1.0 |
AugMix + P-Shuffle / L2 | ImageNet-1k | Pre-train | 1.0 |
AugMix + P-Rotate / L2 | ImageNet-1k | Pre-train | 1.0 |
AugMix + P-Infill / L2 | ImageNet-1k | Pre-train | 1.0 |
AugMix + P-Shuffle / Contrastive | ImageNet-1k | Pre-train | 1.0 |
Rand-Augment + P-Shuffle / Uniform | ImageNet-21k | Pre-train | 0.5 |
Rand-Augment + P-Shuffle / L2 | ImageNet-21k | Pre-train | 0.5 |
Rand-Augment + P-Shuffle / Contrastive | ImageNet-21k | Pre-train | 0.5 |
Rand-Augment + P-Rotate / Uniform | ImageNet-1k | Fine-tune | 0.5 |
Rand-Augment + P-Infill / Uniform | ImageNet-1k | Fine-tune | 1.0 |
AugMix + P-Rotate / Uniform | ImageNet-1k | Fine-tune | 1.0 |
Rand-Augment + P-Shuffle / Uniform | ImageNet-21k | Fine-tune | 0.5 |
Image Size | Stage | Model | Transformation | Hyperparameter |
---|---|---|---|---|
Pre-train | ViT-B/32 | P-Rotate | patch size = 32 | |
Pre-train | ViT-B/16 | P-Shuffle | patch size = 32 | |
Pre-train | ViT-B/16 | P-Rotate | patch size = 16 | |
Pre-train | ViT-B/16 | P-Infill | replace rate = 15/49 | |
Fine-tune | ViT-B/32 | P-Rotate | patch size = 64 | |
Fine-tune | ViT-B/16 | P-Shuffle | patch size = 64 | |
Fine-tune | ViT-B/16 | P-Rotate | patch size = 32 | |
Fine-tune | ViT-B/16 | P-Infill | replace rate = 3/8 |
Appendix C Hyper-parameters in Patch-based Negative Augmentation
For the temperature used in contrastive loss, we consistently observe that works better in pre-training stage and works better in fine-tuning stage. Therefore, we keep this setting for all the models in our paper.
Since we sweep the coefficient in Eqn. 1 from the set , we observe that for most of the cases, works the best. In total we have 48 models using loss regularization on negative views in Table 1, Table 2, Table 5 and Table 3. We use for all of them except those listed in Table 8, where either or works better. Actually, we find our proposed negative augmentation is relatively robust to . Therefore, we suggest using if readers do not want to sweep for the best value for this hyperparameter.
In Table. 9, we display the hyperparameters in each patch-based transformation that we use for the reported results in this work. Our algorithms are generally insensitive to these parameters, and we use the same hyperparameter for all the settings investigated in this work.
Model | CIFAR-100 | CIFAR100-C |
---|---|---|
ViT-B/16 | 91.8 | 74.6 |
+ P-Shuffle / Uniform | 92.6 (+0.8) | 77.0 (+2.4) |
Model | =24 | =32 | ImageNet-1k |
---|---|---|---|
ViT-B/16 () | 32.9 | 57.8 | 84.1 |
BiT-ResNet101-3 | 9.2 | 24.6 | 84.0 |
Model | ImageNet-C Corruption Severity Level | ||||
---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | |
Rand-Augment (Cubuk et al., 2020) | 70.4 | 63.7 | 57.9 | 48.2 | 36.1 |
+ P-Shuffle / Uniform | 71.0 (+0.6) | 64.4 (+0.7) | 59.0 (+1.1) | 49.5 (+1.3) | 37.3 (+1.2) |
+ P-Rotate / Uniform | 71.1 (+0.7) | 64.6 (+0.9) | 59.0 (+1.1) | 50.0 (+1.8) | 37.6 (+1.5) |
+ P-Infill / Uniform | 71.1 (+0.7) | 64.6 (+0.9) | 59.1 (+1.2) | 49.5 (+1.3) | 37.3 (+1.2) |
+ P-Shuffle / L2 | 70.5 (+0.1) | 63.9 (+0.2) | 58.3 (+0.4) | 48.6 (+0.4) | 36.6 (+0.5) |
+ P-Rotate / L2 | 71.1 (+0.7) | 64.8 (+1.1) | 59.5 (+1.6) | 50.1 (+1.9) | 37.8 (+1.7) |
+ P-Infill / L2 | 70.5 (+0.1) | 63.8 (+0.1) | 58.2 (+0.3) | 48.4 (+0.2) | 36.0 (-0.1) |
AugMix (Hendrycks et al., 2020) | 71.4 | 65.2 | 60.5 | 51.9 | 40.2 |
+ P-Shuffle / Uniform | 71.6 (+0.2) | 65.7 (+0.5) | 61.2 (+0.7) | 52.8 (+0.9) | 41.4 (+1.2) |
+ P-Rotate / Uniform | 71.7 (+0.3) | 65.7 (+0.5) | 61.1 (+0.6) | 52.7 (+0.8) | 41.4 (+1.2) |
+ P-Infill / Uniform | 71.9 (+0.5) | 65.8 (+0.6) | 61.1 (+0.6) | 52.4 (+0.5) | 40.8 (+0.6) |
+ P-Shuffle / L2 | 71.8 (+0.4) | 65.8 (+0.6) | 61.0 (+0.5) | 52.4 (+0.5) | 40.7(+0.5) |
+ P-Rotate / L2 | 71.9 (+0.5) | 66.0 (+0.8) | 61.5 (+1.0) | 52.9 (+1.0) | 41.6 (+1.4) |
+ P-Infill / L2 | 71.8 (+0.4) | 65.8 (+0.6) | 61.3 (+0.8) | 52.7 (+0.8) | 41.0 (+0.8) |
Appendix D Ablation Study
D.1 Sensitivity analysis
We test the sensitivity of our patch-based negative augmentation to various patch sizes in P-Shuffle and P-Rotate, and different replace rates in P-Infill. We find that P-Shuffle and P-Rotate are insensitive to patch sizes from for ViT-B/16, and P-Infill is robust to replace rates ranging from 1/3 to 1/2. The accuracy difference is smaller than 0.5 on ImageNet-1k as well as ImageNet-A and ImageNet-R. Therefore, we use the same parameter for all the settings investigated in this work (see Table 9 and Appendix C for details).
D.2 Double batch-size of baselines
As we use the negative augmented view per example, the effective batch size is doubled compared to the vanilla ViT-B/16 trained with only cross-entropy loss. Therefore, we further investigate if the robustness improvement is a result from a larger batch size. When we increase the batch size from 4096 to 8192 in pre-training while keeping the same 300 training epochs, it decreases the in-distribution accuracy to 76.0 on ImageNet-1k as well as the accuracy on robustness benchmarks, e.g., ImageNet-R from 20.3 to 19.3. Hence we conclude the robustness improvement is from the negative data augmentation.
D.3 Pre-training vs. Fine-tuning
We further disentangle the effect of patch-based negative data augmentation in pre-training and fine-tuning. Take P-Shuffle as an example, we design experiments to apply negative augmentation 1) only at the fine-tuning stage, 2) only at the pre-training stage, and 3) at both stages. As shown in Table 13, compared to the baselines, patch-based negative augmentation can effectively help improve robustness in both stages, and its effect in pre-training is slightly larger than in fine-tuning. Finally, we found using negative augmentation in both stages during training yields the largest gain.
Pre-train on ImageNet-1k | |||||
---|---|---|---|---|---|
Model | Stage | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
Rand-Augment (Cubuk et al., 2020) | - | 79.1 | 7.2 | 55.2 | 23.0 |
+ P-Shuffle / Uniform | Fine-tune | 79.1 | 7.1 | 55.3 | 23.0 |
+ P-Shuffle / Uniform | Pre-train | 79.3 | 7.6 | 56.2 | 23.5 |
+ P-Shuffle / Uniform | Both | 79.3 | 7.7 | 56.2 | 23.4 |
+ P-Shuffle / Contrastive | Fine-tune | 79.5 | 7.6 | 56.2 | 23.7 |
+ P-Shuffle / Contrastive | Pre-train | 79.4 | 8.5 | 56.8 | 24.0 |
+ P-Shuffle / Contrastive | Both | 79.7 | 8.9 | 57.8 | 24.7 |
Pre-train on ImageNet-21k | |||||
Model | Stage | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
Rand-Augment (Cubuk et al., 2020) | - | 84.4 | 28.7 | 67.2 | 38.7 |
+ P-Shuffle / L2 | Fine-tune | 84.5 | 29.4 | 67.9 | 39.0 |
+ P-Shuffle / L2 | Pre-train | 84.4 | 29.9 | 67.5 | 38.8 |
+ P-Shuffle / L2 | Both | 84.5 | 29.7 | 68.0 | 39.6 |
+ P-Shuffle / Contrastive | Fine-tune | 84.4 | 29.2 | 67.5 | 38.7 |
+ P-Shuffle / Contrastive | Pre-train | 84.6 | 29.9 | 67.7 | 38.5 |
+ P-Shuffle / Contrastive | Both | 84.3 | 30.8 | 68.1 | 38.6 |
Pre-train on ImageNet-1k | ||||
---|---|---|---|---|
Model | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
ViT-B/16 + Contrastive* | 78.7 | 8.1 | 53.5 | 22.8 |
ViT-B/16 + Shuffle / Contrastive | 78.9 | 8.2 | 54.1 | 23.2 |
ViT-B/16 + P-Rotate / Contrastive | 78.9 | 8.6 | 54.1 | 23.6 |
Rand-Augment + Contrastive* | 79.7 | 8.9 | 57.6 | 24.7 |
Rand-Augment + P-Rotate / Contrastive | 79.9 | 9.4 | 58.4 | 25.4 |
Rand-Augment + P-Infill / Contrastive | 79.9 | 9.3 | 57.9 | 25.0 |
AugMix + Contrastive* | 79.6 | 9.0 | 59.8 | 27.2 |
AugMix + P-Rotate / Contrastive | 79.6 | 9.8 | 60.0 | 27.5 |
AugMix + P-Infill / Contrastive | 79.6 | 9.9 | 60.3 | 27.3 |
Pre-train on ImageNet-21k | ||||
Model | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
Rand-Augment + Contrastive* | 84.1 | 29.7 | 67.6 | 39.2 |
Rand-Augment + P-Shuffle / Contrastive | 84.3 | 30.8 | 68.1 | 38.6 |
D.4 When to use positive data augmentation
As Steiner et al. (2021) observed that traditional (positive) augmentation can slightly hurt the accuracy of ViT if applied to fine-tuning stage, we compare the accuracy of a ViT-B/16 when positive augmentation (e.g., Rand-Augment (Cubuk et al., 2020)) is only applied to pre-training stage as well as both stages. As shown in Table 15, fine-tuning without Rand-Augment achieves slightly better performance.
Model | Stage | ImageNet-1k | ImageNet-A | ImageNet-C | ImageNet-R |
---|---|---|---|---|---|
Rand-Augment | Pre-train | 84.4 | 28.7 | 67.2 | 38.7 |
Rand-Augment | Both | 84.4 | 29.1 | 67.0 | 38.4 |
Appendix E Visualization of Patch-based Transformations
We display more examples with patch-based transformations without cherry-picking in Figure 6, Figure 7 and Figure 8.



































