DATA: Domain-Aware and Task-Aware Self-supervised Learning
Abstract
The paradigm of training models on massive data without label through self-supervised learning (SSL) and finetuning on many downstream tasks has become a trend recently. However, due to the high training costs and the unconsciousness of downstream usages, most self-supervised learning methods lack the capability to correspond to the diversities of downstream scenarios, as there are various data domains, different vision tasks and latency constraints on models. Neural architecture search (NAS) is one universally acknowledged fashion to conquer the issues above, but applying NAS on SSL seems impossible as there is no label or metric provided for judging model selection. In this paper, we present DATA, a simple yet effective NAS approach specialized for SSL that provides Domain-Aware and Task-Aware pre-training. Specifically, we (i) train a supernet which could be deemed as a set of millions of networks covering a wide range of model scales without any label, (ii) propose a flexible searching mechanism compatible with SSL that enables finding networks of different computation costs, for various downstream vision tasks and data domains without explicit metric provided. Instantiated With MoCo v2, our method achieves promising results across a wide range of computation costs on downstream tasks, including image classification, object detection and semantic segmentation. DATA is orthogonal to most existing SSL methods and endows them the ability of customization on downstream needs. Extensive experiments on other SSL methods demonstrate the generalizability of the proposed method. Code is released at https://github.com/GAIA-vision/GAIA-ssl.
1 Introduction

As is universally agreed that deep learning algorithms are data-hungry, how to leverage the exponentially growing unlabelled data from open-source has become a huge challenge. Self-supervised learning (SSL), which utilizes the inherent relationships of data to formulate supervision without manual annotation, has made remarkable progresses in both natural language processing (NLP) [17, 14, 33, 34] and computer vision (CV) [19, 18, 10] area. Despite its great success, to unleash the true power of SSL requires gigantic scale of data and unimaginable training budgets. This brings about a side-effect that it is extremely expensive to train models of various architectures to cover the heterogeneous downstream needs, considering most scenarios are requesting for models of different scales and different downstream vision tasks may desire different model architectures. It is usually believed that neural architecture search (NAS) is designed for solving the issues above. However, labels are so indispensable for existing NAS methods that it seems impossible to apply NAS in SSL, because there is no clue for model selection if no label or metric is provided. These reflections leave us two problems:
(1) Is it possible to train networks of distinctive architectures simultaneously in SSL? Gladly, previous methods [43, 5, 4, 8] have proved that training a supernet that comprises of millions of weight-sharing subnets is possible in the regime of supervised learning. Thus the hardship only lies in how to prevent the co-training of different networks from diverging when there is no strong and stable supervision. Some recent studies [8, 48] interpret the process of SSL of siamese based methods as a form of self-distillation that the query-branch works as a student and the teacher-branch works as a teacher. Thus if we stabilize the behaviour of the teacher, we could provide a relatively steady knowledge source for the heterogeneous students. In this work, we build a supernet training mechanism for siamese based SSL that we fix the key-branch with the maximum architecture of supernet as a teacher and vary only the architectures of the query-branch. Experiments show that this ensures the efficiency of convergence and greatly improves the capability of feature representation of small subnets. More importantly, this design of training supernet in SSL brings us the answer to the critical question below.
(2) How to judge the quality of a network if no label or metric is provided? It is generally agreed that the bigger the better works for deep neural networks when data is sufficient. Given a supernet that covers subnets of different sizes and the knowledge distillation behaviour of SSL, the distance between subnets and the maximum network naturally becomes a self-supervised metric for judging the quality of networks. This metric works well especially when there is a budget constraint for subnets. More discussions about this assumption are placed in Sec 6
We further extend our exploration to enable the searching process to be aware of the type of downstream tasks that different tasks adopt different types of features for measuring the distance of student and teacher. This greatly minimizes the gap transferring to downstream tasks while keeping the searching strategy plug-and-play.
As shown in Figure 1, our approach enables training models of various sizes all in one go and searching appropriate models specialized for specific downstream tasks, computation constraints and data domains. This entire pipeline does not require any label for training or model selection. Instantiated with MoCo v2 [11] , we validate our contributions on evaluating our models on several standard of self-supervised benchmarks. We also combine our approach with other existing SSL methods [18, 48, 37] to demonstrate the generalizability.
2 Related Work
2.1 Self-supervised learning
Self-supervised learning has become the main paradigm of unsupervised learning. It aims at building a good pretext to learn fruitful feature representation from data itself. These pretexts can be mainly separated by two categories: reconstruction-based which concludes colorization [47], spatial jigsaw puzzles [38], inpainting [44] and discriminant-based which contains rotation predict [25], instance-level contrastive [19, 10, 18], and fine-grained contrastive [37, 41].
Contrastive learning.
For the first time, contrastive methods [19, 10] makes self-supervised training become comparable with supervised counterpart. Their method mainly focus on pulling representations of different views of the same image (positive pairs) closer and pushing representations of different images (negative pairs) away in the same time. Further, [7, 18] just use positive pairs to make network learn fruitful feature. These aforementioned methods pretext are sub-optimal in some extent for dense predict task (object detection, semantic segmentation, etc.). To relive this issue, fine-grained pretext [37, 41, 30, 2, 23] are proposed. Most of these methods already outperform supervised counterpart in some dense predict downstream tasks.
2.2 Neural Architecture Search
Neural architecture search aims at automating the architecture design process under certain constraints. [49, 50] proposes to use reinforcement learning with the metric on proxy datasets as the reward for solving this problem. But due to the unaffordable cost, one-stage NAS [29, 6, 15, 1, 3] are proposed, which train and search candidate architectures inside a single supernet. Though getting a specific architecture easily with those methods, we still need to train and search from scratch to get a new architecture once the constraints (such as latency, memory cost) changed. Further, researchers propose methods [45, 43, 5, 4, 9] to achieve training one supernet, which can contain a series of subnets that cover a wide range of scenarios. Their effective supervised training methodology inspired us the most. In [28], it firstly shows that network architectures performance on self-supervised tasks like rotation prediction is linearly correlated with the performance on the supervised task which inspired the design of searching mechanism in our method.
3 Method
3.1 Preliminaries
We formulate the process of common siamese based SSL as a process of dynamic knowledge distillation, which shares the same notion with [8]. To be convenient, we alias the query-branch as student branch, and the key-branch as teacher branch in our paper. Given unlabeled samples , two views ( and ) are obtained on each sample through composition of different augmentations and fed into a student and a teacher network , parameterized by and respectively. In most cases, the teacher shares the exponential moving averaged (EMA) weights of student, namely, . We use and to denote the encoded features from student and teacher models, respectively. is used to represent the similarity function. Taking MoCo [19] as an example, the InfoNCE loss [32] (Eq. 1) is adopted for training model:
(1) |
where is a temperature hyper-parameter [40].
3.2 Self-supervised Supernet Training
No single model could perfectly match the needs of heterogeneous downstream applications, as there might be different latency constraints, data domains and task gaps. Thus we aim to train a great deal of models together instead of training a single one in the regime of SSL, and we hope they cover a wide range of model scale. We extend the definition of a network from to , with a new dimension meaning model architecture. Concretely, we build up a supernet which contains numerous weight-sharing [45] subnets of various architectures , formulated as:
(2) |
where is the total number of subnets. Particularly, we mark the largest model in supernet as , because weights of all the subnets are completely included by it. During training, we fix the architecture of teacher-branch as the model- which means , and vary the architecture of student-branch as . The weights of teacher-branch use the EMA version of the model-, namely, .
In each training iteration, we random sample two network architectures from together with the maximum one to form a architecture set . Following the conventional training regime in siamese-based SSL, we feed to teacher-branch and generate the embedded feature of teacher , and we feed to the models in student-branch from to get student features where . With similarity measured by dot product, we apply InfoNCE loss on , back-propagate the gradients and update all the involved parameters. The whole training process is shown as pseudo-code in Algorithm 1
Model space of supernet.
We choose the popular ResNet [21] as the basic architecture in our work. Depth111Number of blocks in each stage. and width222Number of channels of convolutions in each stage. are adopted as factors to formulate the model space. Sharing the same notion with [21], the output feature maps of each stage are denoted as for future use. As shown in Table 1, the depth of stage start from (2, 2, 5, 2) to (4, 6, 29, 4) with a step of (1, 2, 2, 1), and the width of and each stage start from (32, 48, 96, 192, 384) to (64, 80, 160, 320, 640) with a step of (16, 16, 32, 64, 128).
layer name | ||||
---|---|---|---|---|
[32, 64] | 16 | - | - | |
[48, 80] | 16 | [2, 4] | 1 | |
[96, 160] | 32 | [2, 6] | 2 | |
[192, 320] | 64 | [5, 29] | 2 | |
[384, 640] | 128 | [2, 4] | 1 |

3.3 Self-supervised Model Selection
This part reveals the core motivation behind our adoption of supernet in the regime of SSL. Beyond the value of training massive models simultaneously, it presents a feasible metric for self-supervised NAS and provides abundant architecture candidates for searching. As is universally acknowledged that the bigger the better works mostly for deep neural networks, the distance between subnets and the largest model naturally becomes a metric for model selection.
Domain-awareness.
This mechanism enables us to apply NAS on the downstream data during self-supervised learning, which shrinks the domain gap. Specifically, given a constraint of computation budget , we feed the downstream data into teacher branch and obtain the exemplar feature representations for . Then we randomly sample subnets under the constraint of budget and collect feature representations for each model on the downstream data. We denote the similarity function by and the conquering model is selected to maximize the similarity on the entire downstream dataset:
(3) |
The whole process does not involve any fully-supervised metric such as accuracy or precision.
Task-aware metrics.
The gaps between different vision tasks are always unignorable. To enable our pre-trained supernet to serve various downstream vision tasks, the architecture search has to be conducted under task-aware metrics. Model selected under a single self-supervised metric is not suitable for different tasks. For instance, classifier of image classification task mostly handles the - feature that focuses on the global information of image, while in object detection, equipped with FPN [26], detectors use the multi-stage features of backbone for inference. These task-specific types of features matter most for the specific downstream task. Hence we search for different models by measuring the distance of the features directly used in head of downstream task. We call this task-aware metrics. The pipeline of our method is demonstrated in Figure 2.
For task of image classification, we direct use the feature of . For object detection, we adopt feature of C5 for Faster-RCNN-C4 and features from C2-C5 for Faster-RCNN-FPN. And for task of semantic segmentation, we adopt features from C4-C5. The influence of task-aware metric is detailed in Table 9.
Similarity of relative relations.
During the task-aware searching, the size of the feature map of subnets might be inconsistent with it of teacher. Thus we turn to utilize the relative relations of features to evaluate the similarity of two feature maps. We use and to represent features from student and teacher respectively, where , denote the height and width, and represent number of channels, and means the feature in each column of matrix . We define to express the vector relative relation of vector and as in Eq. 4, and the similarity of relative relation on feature maps can be formulated in Eq. 5. We combine this with the task-aware metric, and conduct model selection in Eq. 3.
(4) |
(5) |
4 Experiments
4.1 Experiments Setup
Datasets. We instantiate our method with the popular MoCo v2 [11] and train on ImageNet [13] which has 1.28 million images in 1000 categories. For the purpose of verifying transferability of models we search on downstream tasks, we experiment on ImageNet [13] semi-supervised classification, COCO [27] instance segmentation, PASCAL VOC [16] object detection, and Cityscapes [12] semantic segmentation. In the ablation study, to validate the generalizability of our method, we combine our method with other self-supervised learning methods [18, 48, 37] and train them on ImageNet-.
Training details.
When supernet trained on Imagenet, we use SGD as the optimizer. The SGD weight decay is 0.000075. We train 200 epochs using a batch size of 1024 on 16 GPUs and an initial learning rate of 0.12. For supernet trained on ImageNet-, we all follow their official default settings.
Searching details.
In order to verify that our method can effectively deal with various scenarios, we compute FLOPs for all of our subnets according to 224x224 input resolution and divide them into seven groups at 1 GFLOPs interval from 1 GFLOPs 8 GFLOPs. We randomly sample one hundred subnets in different groups and search the network in target dataset according to the task-aware metric to find the best one in every group. When verifying our method on various downstream tasks, we will display our experiment results of each group.
4.2 Ranking Correlation
Firstly, we verify that the similarity of the task-specific features between pre-trained subnets and the largest network can be a reliable indicator to assess the performance of these subnets. We sample fifty subnets uniformly and experiment on the above ImageNet- (IN-) semi-supervised classification, VOC object detection, COCO instance segmentation and Cityscapes semantic segmentation. We compute the Spearman [36] correlation between the similarity ranking and their final performance ranking. The results are shown in Table 2. For the low ranking correlation of semantic segmentation, we infer the reason is that the data distribution of the Cityscapes [12] dataset is largely different from the ImageNet [13] dataset used in the self-supervised training stage.
Architecture | Dataset | Feature | Correlation |
---|---|---|---|
ResNet-FC [21] | IN-[13] | 0.90 | |
FasterRCNN-C4 [35] | VOC [16] | C5 | 0.84 |
MaskRCNN-FPN [20] | COCO [27] | C2-C5 | 0.86 |
FCN [31] | Cityscapes [12] | C4-C5 | 0.63 |
4.3 Results on Various Downstream Tasks
Linear evaluation on ImageNet. We train supervised linear classifiers (a fully-connected layer with softmax) on frozen features after BN statics calibration for all of networks we search, following the procedure described in [42] (detailed in supplementary materials). We report 1-crop, top-1 classification accuracy on the ImageNet validation set. The result is shown in Figure 3 and the network architectures we search are described in appendix. We notice .The top-1 accuracy of linear evaluation of the model we search in Group 3G-4G is which outperforms ResNet50 (3.8G, ) by .

Model | FLOPs | Params | Depth | Width | ||||||
---|---|---|---|---|---|---|---|---|---|---|
R50* [11] | 3.8G | 25.5M | [3, 4, 6, 3] | [64, 64, 128, 256, 512] | 38.7 | 59.2 | 42.3 | 35.5 | 56.2 | 37.9 |
R50 [39] | 3.8G | 25.5M | [3, 4, 6, 3] | [64, 64, 128, 256, 512] | 38.6 | 59.5 | 42.1 | 35.2 | 56.3 | 37.5 |
Group | FLOPs | Params | Depth | Width | ||||||
1G2G | 1.8G | 13.6M | [2, 2, 5, 2] | [32, 48, 96, 192, 512] | 36.2 | 56.6 | 39.5 | 33.4 | 53.9 | 35.4 |
2G3G | 2.7G | 14.7M | [2, 2, 13, 2] | [48, 48, 96, 192, 384] | 38.3 | 58.4 | 41.8 | 34.8 | 55.2 | 37.4 |
3G4G | 3.7G | 25.7M | [3, 2, 17, 3] | [32, 48, 96, 192, 512] | 39.9 | 60.2 | 43.5 | 36.0 | 57.1 | 38.6 |
4G5G | 4.2G | 33.1M | [2, 2, 25, 2] | [64, 64, 128, 192, 384] | 40.4 | 60.9 | 44.3 | 36.5 | 58.0 | 39.3 |
5G6G | 5.9G | 43.4M | [4, 6, 21, 4] | [32, 64, 96, 192, 640] | 41.2 | 61.9 | 45.2 | 37.2 | 58.5 | 39.9 |
6G7G | 6.6G | 40.2M | [3, 6, 27, 3] | [64, 80, 96, 192, 640] | 41.5 | 62.1 | 45.1 | 37.3 | 58.9 | 40.1 |
We find that relative small models benefit most from the training strategy. But the searched architecture in 7G8G group performs slightly worse than R101. We infer that when one model architecture outperforms the largest one, its output features are also far away from the largest teacher.
Semi-supervised classification on ImageNet.
Next, following [10], we evaluate performance of the architectures we search on ImageNet set after fine-tuning on and subset of ImageNet set with annotations. For the convenience of description, we use ImageNet- and ImageNet- to represent the two subsets. The training procedure is detailed in supplementary materials. The results of Top-1 on the set are reported in Table 4. We obtain boost when fine-tuned on ImageNet- and gain on ImageNet-. For ImageNet- semi-supervised setting, even the model selected from 1G2G group outperforms the ResNet50 baseline by a large margin.
Model | Params | Top-1 | Top-5 | ||
---|---|---|---|---|---|
R50* [11] | 25.5M | 39.8 | 61.8 | 68.3 | 85.1 |
Group | Params | ||||
1G2G | 14.7M | 45.8 | 61.9 | 73.7 | 84.9 |
2G3G | 19.5M | 46.5 | 63.0 | 74.7 | 85.8 |
3G4G | 33.5M | 47.5 | 63.4 | 75.1 | 85.7 |
4G5G | 37.0M | 47.8 | 64.3 | 75.3 | 86.4 |
5G6G | 43.4M | 49.0 | 65.1 | 76.2 | 86.7 |
6G7G | 45.4M | 49.3 | 65.4 | 76.3 | 87.0 |
7G8G | 45.2M | 49.5 | 65.6 | 76.5 | 86.9 |
COCO instance segmentation.
For COCO instance segmentation, we follow the common setting [19] that fine-tune a Mask R-CNN detector (FPN) on COCO split (118k images) for all the architectures of our search with standard schedule and evaluating on COCO split. The results are shown in Table 3. The pre-train architecture under 4 GFLOPs of our search outperforms the standard ResNet50 by in detection, and the promotion is for segmentation.
Model | FLOPs | Params | |||
R50[11] | 3.8G | 25.5M | 53.5 | 81.3 | 58.8 |
R50*[11] | 3.8G | 25.5M | 57.2 | 82.4 | 63.7 |
Group | FLOPs | Params | |||
1G2G | 1.8G | 12.2M | 50.9 | 79.8 | 55.2 |
2G3G | 2.8G | 25.0M | 58.0 | 82.7 | 64.5 |
3G4G | 3.7G | 25.6M | 58.5 | 83.0 | 65.0 |
4G5G | 4.2G | 33.1M | 59.1 | 83.2 | 65.3 |
5G6G | 5.9G | 43.4M | 60.5 | 83.9 | 67.1 |
6G7G | 6.9G | 47.4M | 60.4 | 83.5 | 67.0 |
7G8G | 7.3G | 45.3M | 60.4 | 83.6 | 67.3 |
Object detection on PASCAL VOC.
When transferring to VOC [16] object detection, following the [11]: a Faster R-CNN detector [35] (C4-backbone) is adopted. And it is trained on VOC 07+12 set for 24K iterations and evaluated on the VOC 2007 set. The results are presented in Table 5. With our framework, we can provide a model whose FLOPs and Params are similar to standard ResNet50, outperforms the R50 baseline by 1.2 AP.

Model | Depth | Width | mIoU |
---|---|---|---|
R50 | [3, 4, 6, 3] | [64, 64, 128, 256, 512] | 75.5 |
R50* [11] | [3, 4, 6, 3] | [64, 64, 128, 256, 512] | 76.4 |
Group | Depth | Width | mIoU |
1G2G | [2, 2, 5, 2] | [48, 48, 96, 192, 512] | 72.7 |
2G3G | [3, 2, 9, 2] | [48, 48, 96, 192, 512] | 75.2 |
3G4G | [2, 2, 17, 3] | [32, 48, 96, 192, 384] | 77.4 |
4G5G | [2, 6, 19, 3] | [32, 48, 96, 192, 640] | 77.0 |
5G6G | [2, 4, 25, 3] | [64, 64, 128, 192, 640] | 78.1 |
6G7G | [4, 6, 23, 4] | [32, 64, 128, 192, 640] | 77.6 |
7G8G | [4, 6, 21, 4] | [32, 64, 160, 192, 640] | 78.2 |
Cityscapes semantic segmentation.
Cityscapes [12] is a widely used benchmark for semantic segmentation. Following [37], we fine-tune the backbone of our search in the FCN [31] form on set (2975 images) for 40k iterations with batch size 16 and test on set (500 images). The results are reported in Table 6. Although the correlation of our ranking strategy is fair, as shown in Table 2, we can still find the network outperforms the baseline with our effective training by 1 mIoU.
Transfer to other classification tasks.
Furthermore, we evaluate our framework on more diverse classification datasets in VTAB [46] (detailed in supplementary materials). We only find the most similar pre-train architecture in 3G4G FLOPs group for the comparison with standard ResNet50. We perform fine-tuning on these datasets and report the results in Figure 4. The training details are detailed in supplementary materials. Compared to our MoCo v2 [19] baseline, we can get consistent improvement on these datasets. Although MoCo v2 is at a disadvantage in classification compared to supervised pre-training, thanks to the advantages of our training strategy and model customization, it can exceed the performance of supervised pre-training on some datasets. As the results show, the performance on eurasat [22] dataset attracts our attention. This dataset is a land cover classification dataset, and we find its data distribution is completely different from ImageNet [13]. Hence, when transferring to this dataset, these feature representations are of no use.
5 Ablation Study
Combining with contrastive learning methods. Our proposed pipeline is orthogonal to most contrastive learning methods. The reason for choosing MoCo v2 as baseline lies in its comprehensive performance on various downstream tasks. To demonstrate the generalizability of DATA, we instantiated with three classic contrastive self-supervised learning methods, BYOL [18], ReSSL [48] and DenseCL [37]. Due to the huge computation cost for training self-supervised models, we conduct experiments of these methods only on ImageNet-. Since [18, 48] are designed for classification task originally and [37] is mainly designed for the dense-prediction task, we report comparison results on area of their expertise. As shown in Table 7, consistent improvement could be observed.
Method | Task | performance |
---|---|---|
BYOL (repro) | Cls | 23.7 |
Ours+BYOL | Cls | 24.8 |
ReSSL (repro) | Cls | 23.7 |
Ours+ReSSL | Cls | 26.7 |
DenseCL (repro) | Det | 49.1 |
Ours+DenseCL | Det | 50.2 |
Ablation with teacher architecture.
In this ablation study, we explore the impact of choice on teacher architecture. Instead of fixing the architecture of teacher branch to the largest network, we compare with the setting using the same architecture as the sampled subnet in student branch, namely, . We base the experiment on MoCo v2 and train supernets on ImageNet-. Next, we extract the standard ResNet50 from supernet and evaluate it on the val set following the setting of linear classification. As reported in Table 8, we see that fixing the architecture of teacher branch is crucial, that the top-1 accuracy reaches under linear evaluation protocol, higher than the unfixed setting. This convey a message that a stable teacher is important for self-supervised supernet training. We also observe that this result outperform the vanilla MoCo v2 by , which means supernet distillation is helpful over self-distillation.
Method | Fixed teacher arch | Top-1 | Top-5 |
---|---|---|---|
MoCo v2 | |||
Ours | |||
Ours | ✓ |
Ablation with task-aware metrics. Here we explore the effectiveness of task-aware metrics for model selection. For each downstream task, we select models according to metrics based on , C5 and C2-C5 respectively.
As reported in Table 9, task-aware metrics yield , and improvements when task and metric are matched.
Dataset | Task | Architecture | Feature | Performance |
---|---|---|---|---|
47.5 | ||||
IN- [13] | Semi-cls | ResNet-FC | C5 | 47.3 |
C2-C5 | 47.1 | |||
57.8 | ||||
VOC [16] | Det | FasterRCNN-C4 | C5 | 58.5 |
C2-C5 | 58.1 | |||
39.2 | ||||
COCO [27] | Det | MaskRCNN-FPN | C5 | 39.3 |
C2-C5 | 39.9 |
Ablation with domain-awareness.
We also explore the influence of domain-awareness.
Specifically, we compare our searched models above with models searched through different datasets. Results are reported in Table 9. Note that these models are all in 3G4G group.
We find that searching by ImageNet seems to be an acceptable indicator for the performance of object detection on COCO. While in task of segmentation, searching without domain-awareness severely degrades the correlation() and mIoU().
Task | Target | Source | Correlation | Performance |
---|---|---|---|---|
Det | COCO | ImageNet | 0.82 | 39.7 |
Det | COCO | COCO | 0.86 | 39.9 |
Seg | Cityscapes | ImageNet | 0.23 | 76.2 |
Seg | Cityscapes | Cityscapes | 0.63 | 77.4 |
6 Limitation
The major limitation of this work lies in the imbalanced training among subnets. Specifically, we find that the smaller two-thirds of subnets in supernet are trained well while the rest larger subnets are not. We infer that subnets could benefit more from knowledge distillation and most of their weights are always covered regardless of which subnet is sampled during training. Conversely, larger subnets have no effective teacher as source of knowledge and their exclusive weights might be under insufficient training.
For our assumption, it indeed can not deal with this situation where candidate subnets are close to the largest teacher network. Because when one model architecture outperforms the largest one, its output features are also far away from the counterpart of largest teacher.
7 Conclusion
We have explored combining NAS with self-supervised learning and positive results are shown. Firstly, we manage to train massive weight-sharing subnets in a supernet simultaneously in the regime of self-supervised learning. More importantly, this mechanism of supernet training makes possible the unlabeled NAS since the feature distance between subnets and the largest network works perfectly as a self-supervised metric for model selection. Our work is orthogonal to most existing self-supervised learning methods and endows them the capability of customization on various downstream needs. We hope our approach could truly useful in real-world applications and our adventure on NAS in SSL could inspire more genius minds.
8 Acknowledgement
We thank Jiawei He, Shuwei Sun, Yuqi Wang, Lin Zhang and anonymous reviewers for their helpful discussions that improved this paper.
This work was supported in part by the Major Project for New Generation of AI (No.2018AAA0100400), the National Natural Science Foundation of China (No. 61836014, No. U21B2042, No. 62072457, No. 62006231).
References
- [1] Youhei Akimoto, Shinichi Shirakawa, Nozomu Yoshinari, Kento Uchida, Shota Saito, and Kouhei Nishida. Adaptive stochastic natural gradient method for one-shot neural architecture search. In ICML, 2019.
- [2] Amir Bar, Xin Wang, Vadim Kantorov, Colorado J Reed, Roei Herzig, Gal Chechik, Anna Rohrbach, Trevor Darrell, and Amir Globerson. Detreg: Unsupervised pretraining with region priors for object detection. arXiv:2106.04550, 2021.
- [3] Andrew Brock, Theodore Lim, James M Ritchie, and Nick Weston. Smash: one-shot model architecture search through hypernetworks. In ICLR, 2018.
- [4] Xingyuan Bu, Junran Peng, Junjie Yan, Tieniu Tan, and Zhaoxiang Zhang. Gaia: A transfer learning system of object detection that fits your needs. In CVPR, 2021.
- [5] Han Cai, Chuang Gan, Tianzhe Wang, Zhekai Zhang, and Song Han. Once-for-all: Train one network and specialize it for efficient deployment. In ICLR, 2020.
- [6] Han Cai, Ligeng Zhu, and Song Han. Proxylessnas: Direct neural architecture search on target task and hardware. In ICLR, 2019.
- [7] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. In NeurIPS, 2020.
- [8] Mathilde Caron, Hugo Touvron, Ishan Misra, Hervé Jégou, Julien Mairal, Piotr Bojanowski, and Armand Joulin. Emerging properties in self-supervised vision transformers. In ICCV, 2021.
- [9] Minghao Chen, Houwen Peng, Jianlong Fu, and Haibin Ling. Autoformer: Searching transformers for visual recognition. In ICCV, 2021.
- [10] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In ICML, 2020.
- [11] Xinlei Chen, Haoqi Fan, Ross Girshick, and Kaiming He. Improved baselines with momentum contrastive learning. arXiv:2003.04297, 2020.
- [12] Marius Cordts, Mohamed Omran, Sebastian Ramos, Timo Rehfeld, Markus Enzweiler, Rodrigo Benenson, Uwe Franke, Stefan Roth, and Bernt Schiele. The cityscapes dataset for semantic urban scene understanding. In CVPR.
- [13] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
- [14] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding. In NAACL, 2019.
- [15] Xuanyi Dong and Yi Yang. Searching for a robust neural architecture in four gpu hours. In CVPR, 2019.
- [16] Mark Everingham, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes (voc) challenge. IJCV, 2010.
- [17] Luciano Floridi and Massimo Chiriatti. Gpt-3: Its nature, scope, limits, and consequences. Minds and Machines, 30(4):681–694, 2020.
- [18] Jean-Bastien Grill, Florian Strub, Florent Altché, Corentin Tallec, Pierre H Richemond, Elena Buchatskaya, Carl Doersch, Bernardo Avila Pires, Zhaohan Daniel Guo, Mohammad Gheshlaghi Azar, et al. Bootstrap your own latent: A new approach to self-supervised learning. In NeurIPS, 2020.
- [19] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In CVPR, 2020.
- [20] Kaiming He, Georgia Gkioxari, Piotr Dollar, and Ross Girshick. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Oct 2017.
- [21] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016.
- [22] Patrick Helber, Benjamin Bischke, Andreas Dengel, and Damian Borth. Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2019.
- [23] Olivier J Hénaff, Skanda Koppula, Jean-Baptiste Alayrac, Aaron van den Oord, Oriol Vinyals, and João Carreira. Efficient visual pretraining with contrastive detection. In ICCV, 2021.
- [24] Haris Iqbal. Harisiqbal88/plotneuralnet v1.0.0. https://doi.org/10.5281/zenodo.2526396, Dec. 2018.
- [25] Nikos Komodakis and Spyros Gidaris. Unsupervised representation learning by predicting image rotations. In ICLR, 2018.
- [26] Tsung-Yi Lin, Piotr Dollár, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In CVPR, 2017.
- [27] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014.
- [28] Chenxi Liu, Piotr Dollár, Kaiming He, Ross Girshick, Alan Yuille, and Saining Xie. Are labels necessary for neural architecture search? In ECCV, 2020.
- [29] Hanxiao Liu, Karen Simonyan, and Yiming Yang. Darts: Differentiable architecture search. In ICLR, 2019.
- [30] Songtao Liu, Zeming Li, and Jian Sun. Self-emd: Self-supervised object detection without imagenet. arXiv:2011.13677, 2020.
- [31] Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In CVPR, 2015.
- [32] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. arXiv:1807.03748, 2018.
- [33] Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. Improving language understanding by generative pre-training. 2018.
- [34] Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
- [35] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster r-cnn: Towards real-time object detection with region proposal networks. In NeurIPS, 2015.
- [36] C Spearman. The proof and measurement of association between two things. The American Journal of Psychology, 15(1):72–101, 1904.
- [37] Xinlong Wang, Rufeng Zhang, Chunhua Shen, Tao Kong, and Lei Li. Dense contrastive learning for self-supervised visual pre-training. In CVPR, 2021.
- [38] Chen Wei, Lingxi Xie, Xutong Ren, Yingda Xia, Chi Su, Jiaying Liu, Qi Tian, and Alan L Yuille. Iterative reorganization with weak spatial constraints: Solving arbitrary jigsaw puzzles for unsupervised representation learning. In CVPR, 2019.
- [39] Yuxin Wu, Alexander Kirillov, Francisco Massa, Wan-Yen Lo, and Ross Girshick. Detectron2. https://github.com/facebookresearch/detectron2, 2019.
- [40] Zhirong Wu, Yuanjun Xiong, Stella X Yu, and Dahua Lin. Unsupervised feature learning via non-parametric instance discrimination. In CVPR, 2018.
- [41] Zhenda Xie, Yutong Lin, Zheng Zhang, Yue Cao, Stephen Lin, and Han Hu. Propagate yourself: Exploring pixel-level consistency for unsupervised visual representation learning. In CVPR, 2021.
- [42] Jiahui Yu and Thomas S Huang. Universally slimmable networks and improved training techniques. In CVPR, 2019.
- [43] Jiahui Yu, Pengchong Jin, Hanxiao Liu, Gabriel Bender, Pieter-Jan Kindermans, Mingxing Tan, Thomas Huang, Xiaodan Song, Ruoming Pang, and Quoc Le. Bignas: Scaling up neural architecture search with big single-stage models. In ECCV, 2020.
- [44] Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, and Thomas S Huang. Generative image inpainting with contextual attention. In CVPR, 2018.
- [45] Jiahui Yu, Linjie Yang, Ning Xu, Jianchao Yang, and Thomas Huang. Slimmable neural networks. In ICLR, 2019.
- [46] Xiaohua Zhai, Joan Puigcerver, Alexander Kolesnikov, Pierre Ruyssen, Carlos Riquelme, Mario Lucic, Josip Djolonga, Andre Susano Pinto, Maxim Neumann, Alexey Dosovitskiy, et al. A large-scale study of representation learning with the visual task adaptation benchmark. arXiv preprint arXiv:1910.04867, 2019.
- [47] Richard Zhang, Phillip Isola, and Alexei A Efros. Colorful image colorization. In ECCV, 2016.
- [48] Mingkai Zheng, Shan You, Fei Wang, Chen Qian, Changshui Zhang, Xiaogang Wang, and Chang Xu. Ressl: Relational self-supervised learning with weak augmentation. In NeurIPS, 2021.
- [49] Barret Zoph and Quoc V Le. Neural architecture search with reinforcement learning. In ICLR, 2017.
- [50] Barret Zoph, Vijay Vasudevan, Jonathon Shlens, and Quoc V Le. Learning transferable architectures for scalable image recognition. In CVPR, 2018.