Open Set Domain Adaptation by Extreme Value Theory
Abstract
Common domain adaptation techniques assume that the source domain and the target domain share an identical label space, which is problematic since when target samples are unlabeled we have no knowledge on whether the two domains share the same label space. When this is not the case, the existing methods fail to perform well because the additional unknown classes are also matched with the source domain during adaptation. In this paper, we tackle the open set domain adaptation problem under the assumption that the source and the target label spaces only partially overlap, and the task becomes when the unknown classes exist, how to detect the target unknown classes and avoid aligning them with the source domain. We propose to utilize an instance-level reweighting strategy for domain adaptation where the weights indicate the likelihood of a sample belonging to known classes and to model the tail of the entropy distribution with Extreme Value Theory for unknown class detection. Experiments on conventional domain adaptation datasets show that the proposed method outperforms the state-of-the-art models.
1 Introduction
Recently, deep learning techniques have drawn a lot of attention in both academia and industry, due to their astounding performance in fields such as computer vision and natural language processing [16, 31, 13, 35, 5, 38, 32, 1, 26, 10]. However, one major disadvantage of deep learning is that neural networks generally require a large amount of training data to converge to the right solution and generalize. When the training data are insufficient, the model performance is usually adversely affected. Sometimes even if the training data are sufficient, the domain gap, i.e. the difference between data distributions, between the source domain (the data we train model on) and the target domain (the desired target task) may still contribute to low generalizability. This is because in conventional machine learning tasks, we usually assume that the training data distribution is the same as the testing data distribution. However, in real world, testing data are uncontrollable, and thus the difference between the source and the target domain can be substantial, which results in the overfitting problem, that is, the model does not generalize well to the testing set.
In order to reduce the domain gap and better utilize the source domain knowledge, domain adaptation techniques have been proposed to resolve the issue. Domain adaptation assumes that the source domain has sufficient amount of labeled data to train a good model, while the desired target domain has insufficient amount of data to train the model. Note that in most domain adaptation tasks, the target domain does not have labeled data at all, which is also known as unsupervised domain adaptation. Domain adaptation methods leverage the knowledge from the source domain with sufficient labeled data to learn a model that works well in the target domain with insufficient or no labeled data. Typically, domain adaptation methods reduce the domain gap by diminishing the divergence between the label-rich source and the label-scarce target domain [14, 15, 11, 12, 36, 25, 34, 30, 21, 8, 7, 28]. A significant drawback of most of the existing works is that they assume the source label space and the target label space are identical, i.e. the target classes are assumed to have appeared during the training process and the training classes do not contain classes that are not in the target domain. This is also known as closed set domain adaptation (CSDA), which is infrequent in real-world scenarios since it is hard to guarantee that the target domain classes are the same as the source domain classes. Aligning the source and the target domain by brute force when their label spaces are different is extremely detrimental to the model’s generalizability, which is also known as the negative transfer phenomenon, because the CSDA methods would try to align the additional unknown classes as well during adaptation.
To handle the domain adaptation tasks without assuming identical label spaces, open set domain adaptation (OSDA) methods were proposed to first detect the irrelevant or unknown samples and then avoid adaptation on the unknown samples and perform domain adaptation only on the known classes [4, 29, 18, 2], which can be achieved by forcing the model to learn a clear boundary between known classes and unknown classes and a clear boundary within the known classes. After adaptation, the model is applied to the target samples as follows: either a target sample is classified as a class among the known classes, or rejected as an unknown class. However, all of the conventional OSDA methods employ a rejection threshold hyperparameter, where if a score or statistic for a sample is higher than the threshold, then the sample will be rejected as an unknown class and discarded during adaptation, and thus the model’s sensitivity is largely affected by this rejection threshold.
In our work, we handle the existing issue in OSDA by 1) utilizing an entropy-based instance-level reweighting strategy and 2) extreme value theory (EVT) which has proven to be useful in many classification tasks due to its ability to model extreme values [9, 23, 6, 37, 3]. We propose to use entropy of probability distribution of a sample to measure the likelihood it belongs to unknown classes. That is, the higher the entropy is, the more likely the sample belongs to the unknown classes, because the model is more uncertain regarding the prediction. We utilize the entropy values to construct an instance-level weight for domain adaptation, instead of setting a hard threshold. In this way, every sample is taken into account during the adaptation process, and a sample with high entropy should be focused less to avoid the negative transfer problem. In inference, we model the tail probability of the entropy distribution by fitting a generalized extreme value (GEV) distribution, and we use the cumulative distribution function (CDF) score to indicate if a sample belongs to unknown or known classes. Experimental results on three conventional domain adaptation datasets show outperformance over both state-of-the-art CSDA and OSDA benchmarks.
We claim the following key contributions. First, we model the tail of entropy distributions with EVT to detect and reject unknown classes. Second, the entropy-based instance-level weighting strategy avoids setting a hard threshold manually and thus is more robust and stable. Third, we have done extensive experiments on three conventional datasets in domain adaptation, which show that our model outperforms benchmarks by a significant margin.
The rest of the paper is organized as follows. In Section 2, we discuss the related literature and methods. In Section 3, we explain our OSDA method in details. In Section 4, we conduct experiments to validate the effectiveness of the proposed method and in Section 5, we conclude the paper by reiterating the main contributions.
2 Related Work
Existing works on CSDA typically try to diminish the domain gap by minimizing a statistical divergence between two domains for adaptation or by an adversarial approach. MCD [28] utilize two task-specific classifiers to detect the target domain samples which are far from the source domain by maximizing the two classifiers’ inconsistency regarding predictions. Based on the mean teacher model [33] that is originally proposed for semi-supervised learning tasks, the self-ensemble network [7] is proposed to calculate the exponentially moving average of the student model weights and it assigns the weights to the teacher model to reduce the domain gap. There are also a plethora of works on adversarial learning to reduce domain gap. DANN [8] reduces the domain gap by introducing a domain discriminator during the training process to discriminate between domains where the domain discriminator is optimized by a specially designed gradient reversal layer. ADDA [34] incorporates adversarial training to reduce the domain gap by the discriminator to distinguish across domains. The soft labeling method [14] utilizes a soft cross-entropy loss function to optimize for the domain invariance and thus aligning the source and target domains. WDGRL [30] is proposed to minimize the Wasserstein distance across domains adversarially for learning domain invariance. CADA [20] aligns the joint distribution for labels and features across two domains by exploring classifier predictions for adversarial domain adaptation. Our proposed method differs from these prior works in that we focus on the practical OSDA setting while these prior works assume an identical label space across domains, and thus they cannot recognize the unknown classes samples included during domain adaptation. Our method is able to detect the unknown classes by instance-level weights based on entropy to avoid negative transfer.
There are also several existing works on OSDA to address the issue of unknown classes. STA [19] is proposed to develop a coarse-to-fine progressive separation method for unknown and known classes. OSBP [29] forces the generator to either match target samples with source known classes or ignore them in adaptation as unknown samples by adversarial training. Based on the work of SE [7] in CSDA, [18] modifies the adaptation loss and then aligns the target domain with the source domain to address the unknown classes using weights calculated by entropy values. By factorization and joint separation, D-FRODA [2] represents the source and target classes with a shared embedding space for domain adaptation. ATI [4] is also proposed to detect the target samples that potentially belong to the known classes by learning a mapping from the source domain to the target domain. We see the difference between our proposed work and the prior works in the utilization of a hard threshold to recognize and reject the unknown samples, and thus the model accuracy largely depends on how the threshold is set. The proposed method utilizes entropy to indicate the likelihood of unknown classes and it constructs a soft instance-level weight. We further fit an EVT model on the tail of the entropy distribution to detect unknown classes. Therefore, by avoiding manual thresholding and introducing EVT to detect unknown classes, the proposed method is more robust and suitable in the OSDA setting, which is also validated in the experimental results.
3 Methodology
In this section, denotes the number of known classes; denote a source, target sample, respectively; denotes a source label; denote the source, target sample distribution, respectively; denote the source, target domain, respectively; denotes a feature extractor; denotes a -way classifier to classify samples into one of the known classes; and denotes the binary domain classifier to classify samples into source or target domains.
Even though all samples in the source domain are labeled, in one loss component we assume a split of known classes to (the classes of interest) and (the irrelevant classes), where , and we treat the classes in as unknown. To this end, we denote by as the sample distribution of the samples corresponding to the classes in (unknown-class distribution). This is a known concept, e.g. [18, 29].
3.1 Loss functions
Entropy Weighted Domain Adversarial Training: In our work, we propose a novel loss to train the model in an adversarial manner by a domain classifier to distinguish samples from the source and target domains, where outputs a probability indicating the likelihood of belonging to the target domain, with instance-level weights on target samples to avoid negative transfer in OSDA. The domain classifier is optimized based on a binary cross-entropy loss. We train the feature extractor to maximize the domain classification loss and the domain classifier to minimize the domain classification loss.
The domain classification loss is calculated by
where
Here is the partitioning function such that .
The feature extractor is trained to maximize so that the embeddings from both domains are similar and the domain classifier is trained to minimize for a good classification performance. Classifier is trained together with a different loss component which is discussed later. We make the feature extractor fool the domain classifier and the domain classifier learns how to perform better from the feature extractor in the adversarial manner.
Entropy Maximization for Source Unknown Classes: The entropy of classifier predictions is high if the sample is from unknown classes and vice versa, because the classifier is optimized to minimize the cross-entropy loss on source known classes. Therefore, the known classes predictions are close to one-hot vectors with low entropy. Based on this observation, the unknown classes detection task can be tackled by separating the known classes and unknown classes by their entropy values. In order to force the classifier to predict unknown classes with high entropy, we penalize the model if the prediction is different from the uniform distribution for a source unknown sample, because we want to force the unknown classes predictions as uncertain as possible. The loss function component is calculated as
By maximizing the source unknown entropy and minimizing the source known entropy, we can further separate the known and unknown classes with a clear boundary based on their entropy values.
Total Loss: Let us denote as the parameters of the feature extractor , as the parameters of the classifier and as the parameters of the domain classifier . The total loss to minimize is calculated as
where is the conventional cross-entropy classification loss on source known classes and are weights for the loss components.
To force the domain classifier to minimize the adversarial loss and the feature extractor to maximize the adversarial loss, we seek a saddle point of satisfying the conditions:
(1) |
On a saddle point, minimize the domain classification loss , minimize the conventional classification loss , maximize the domain classification loss (thus the feature divergence is minimized across the two domains) and minimize the classification loss (thus the features are discriminative).
In the total loss , in order to align domains more effectively in OSDA, we propose a novel loss component with domain discriminator and instance-level weights to reduce the domain gap adversarially. Components and already appear in the existing works such as [18, 29].
Target Sample Classification: We propose to fit GEV using the tail of the entropy distribution of source samples. The probability density function for GEV is calculated as
where
and the CDF is calculated as
where are parameters associated with the distribution.
After fitting GEV using the source samples’ entropy values based on the trained , for target samples, we calculate the CDF based on the source sample GEV. If the CDF value is higher than 0.5, then the sample is classified as an unknown sample; otherwise it is input into the known-class classifier to be predicted into one of the known classes. We abbreviate our overall methodology as ADAGEV (Adversarial Domain Adaptation with GEV).
3.2 Training strategy
We train our model to minimize loss function . Since we want the feature extractor to maximize the adversarial loss, we utilize the gradient reversal layer [8] as the first layer of the domain discriminator such that during forward propagation, the layer is essentially an identity function, while in backward propagation, the gradient is multiplied by . In this way, the total loss function can be jointly optimized in one single step instead of alternate training, i.e. optimizing each loss component individually step by step. There are two advantages of avoiding alternate training: 1. lower computational costs, as fewer gradient iterations are calculated and 2. fewer hyperparameters to tune, e.g. optimizing one component for times and optimizing another component for times are absent, etc.
To estimate the partitioning function, we have tried the following approaches: 1. using the same batch of data (as the batch sampled to calculate the numerator of ) to calculate ; 2. using a different batch of data of the same size to calculate ; 3. using the sum of the values of the first two approaches (sum of weights of the samples). We find that in practice, the first approach achieves overall the best performance with the lowest computational cost, and thus it is used in ADAGEV. The entire algorithm is exhibited in Algorithm 1.
4 Experimental results
We evaluate ADAGEV on three conventional benchmark datasets in domain adaptation: Digits, Office-31 [27] and VisDA [24]. In the experiments, the source domain samples are labeled and the target domain samples are not labeled, and the task is to either classify target samples into one of the known classes or reject as unknown. We use the conventional OSDA metrics OS and OS* as performance measures, which denote the mean accuracy for all classes (all known classes and the additional unknown class) and the mean accuracy for the known classes, respectively. The hyperparameters are set as (discussed later). The benchmark state-of-the-art methods are discussed in Section 2, and the accuracies are obtained from the corresponding papers except [18], which only reported the OS accuracy. To complete the comparison for OS*, we ran the source code of [18] and after finetuning we are able to achieve a slightly better performance for OS than the original paper, and both the OS and OS* accuracies are reported in our experiments. We discuss the network architectures for each experiment in the following subsections. For a fair comparison, the same network architecture is used in ADAGEV as in the benchmark algorithms (which all use the same architecture). The stopping criteria are also discussed in the following subsections.
4.1 Digits Experiments
In the Digits experiments, we follow the conventions in OSDA to use USPS [17], SVHN [22] and MNIST to conduct evaluations, and we perform domain adaptation from SVHN to MNIST, USPS to MNIST and MNIST to USPS. Note that the USPS to MNIST and the MNIST to USPS tasks are easier than the SVHN to MNIST task, since MNIST and USPS both contain 2-dimensional black-and-white images and are hand-written digits, while SVHN contains 3-dimensional RGB images for real-world house numbers which are part of Google Street View images. Therefore, the domain gap between SVHN to MNIST is significantly larger than in the other two experiments. For consistency, we use digits 0, 1, 2, 3 as known classes, 4, 5, 6 as source unknown classes and 7, 8, 9 as target unknown classes, following the conventional protocols. Also following the existing works, we use the same CNN architectures (details can be found in the appendix of [29]) for a fair comparison. We compare ADAGEV with the state-of-the-art methods in both CSDA and OSDA which are discussed in Section 2. We report accuracies for ADAGEV after training for 20 epochs. For benchmarks, they either not mention the stopping criteria or use 200 epoches.
Method | S-M | U-M | M-U | AVG | ||||
---|---|---|---|---|---|---|---|---|
OS | OS* | OS | OS* | OS | OS* | OS | OS* | |
Source only | 58.5 | 63.5 | 82.3 | 83.9 | 82.0 | 84.1 | 74.2 | 77.2 |
DAN [21] | 66.2 | 67.0 | 86.9 | 88.0 | 89.1 | 90.5 | 80.7 | 81.8 |
DANN [8] | 66.8 | 67.4 | 89.2 | 88.9 | 88.9 | 89.9 | 81.6 | 82.0 |
OSBP [29] | 62.2 | 63.9 | 94.8 | 94.3 | 92.7 | 93.2 | 83.2 | 83.8 |
STA [19] | 65.2 | 65.9 | 94.7 | 94.5 | 93.3 | 94.1 | 84.4 | 84.8 |
KASE [18] | 67.2 | 68.1 | 94.7 | 95.1 | 93.6 | 94.5 | 85.2 | 85.9 |
ADAGEV | 87.9 | 89.2 | 95.4 | 96.3 | 95.2 | 96.5 | 92.8 | 94.0 |
Relative Imp | 30.8% | 31.0% | 0.7% | 1.3% | 1.7% | 2.1% | 8.9% | 9.4% |
Table 1 shows the overall performance comparison, where the model trained on source only performs the worst, which is expected because it does not align the two domains at all. The CSDA benchmarks DAN and DANN perform slightly better than training only on source data, because they reduce domain misalignment, but they do not reject unknown classes in domain adaptation. The OSDA methods OSBP, STA and KASE outperform the CSDA methods but only improve by a small margin (1.6%, 2.8%, 3.6%, respectively), from which we observe the difficulty in the OSDA scenario. ADAGEV achieves 92.8% OS accuracy, outperforming the previous state-of-the-art OSDA method by 9.6%, 8.4% and 7.6% respectively. The method also outperforms the CSDA benchmarks by 12.1% and 11.2%. In the most difficult task S-M, our method is able to achieve more than 20% improvement on both OS and OS* accuracies, which again validates the effectiveness of the proposed method in the OSDA scenario. We also experiment on 5 different random seeds and the standard deviations for average OS and OS* are 0.49 and 0.37, indicating the performance is robust and stable. We also observe that comparing the OS and OS* metrics, OS* is always higher than OS for the same method, meaning the unknown class accuracy is lower than the overall accuracy on known classes only. This points to the future research direction in the OSDA area that the unknown classes should be addressed more on their detection and separation. Our relative OS improvements regarding benchmarks are 25.1%, 15.0%, 13.7%, 11.5%, 10.0%, 8.92% which are substantial improvements.
Ablation study In order to have a better understanding on what component in our model contributes most to the performance, we conduct ablation studies by removing components from our model. The ablation study results are exhibited in Figure 1, where only the OS accuracy is shown for simplicity since the OS* accuracy shows a very similar performance, i.e. the OS* accuracy is consistently higher than the OS accuracy by 1% 2%. Note that in the second experiment (i.e. w/o EVT) we replace EVT by a binary classifier to classify target samples into known/unknown classes where the classifier is trained using source known and unknown samples.

From Figure 1 we observe that removing the soft reweighting strategy contributes most to the accuracy drop, i.e. the average OS drops by 10.0% without entropy-based reweighting. Removing the EVT component is also detrimental to the performance, where the average OS drops by 7.3% without EVT. Including both reweighting and EVT in ADAGEV achieves the new state-of-the-art performance in OSDA.
Experiments with less source data In domain adaptation tasks, it is usually assumed that the source domain contains sufficient amount of labeled data. In order to see how our method performs with less source data, we experiment on limiting the source data to 10%, 25% and 50%. From Figure 2, both the OS and OS* accuracies increase when we have more source data. The model trained with 10% data achieves comparable performance with the benchmarks, and the model trained with 25% source data already outperforms the previous state-of-the-art OSDA method. This validates that ADAGEV shows robust performance even when less source data are presented.

Sensitivity on hyperparameters We also experiment on varying the loss function weight hyperparameters to observe their sensitivity. The results are shown in Figure 3, where in general, the hyperparameter for source classification loss is less sensitive in that varying its value results in the least accuracy changes, which we postulate is because the source knowledge can already be learned well for a small weight. The hyperparameters for domain discriminator loss and unknown entropy maximization loss are more sensitive to the changes where either increasing or decreasing the values results in an accuracy drop. The combination which provides the best accuracy is used as the final hyperparameters across all datasets.

4.2 Office-31 Experiments
We also experiment on the Office-31 dataset which is another frequently used domain adaptation dataset containing three domains: Amazon, DSLR and Webcam. The following 6 domain adaptation tasks are created: Amazon to DSLR, DSLR to Amazon, Amazon to Webcam, Webcam to Amazon, DSLR to Webcam and Webcam to DSLR. Each domain contains 31 classes, and the images are either collected directly from www.amazon.com or they are office environment images taken with different lighting, etc. using a webcam or a digital single-lens reflex (DSLR) camera. Therefore, the domain gap between DSLR and Webcam is slightly smaller. Following the same experimental setup as in previous works, we set the 10 common classes from the Office-Caltech dataset [11] as the known classes, and the first 10 remaining classes in the alphabetical order are set as the unknown classes for the source domain and the last 11 classes are set as the unknown classes for the target domain. We use the same CNN model AlexNet [16] as in previous works across all methods for fair comparison. Following the conventions in [29], we report accuracies after training for 50 epoches. The experimental results are presented in Table 2.
Method | A-D | A-W | D-A | D-W | W-A | W-D | AVG | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
OS | OS* | OS | OS* | OS | OS* | OS | OS* | OS | OS* | OS | OS* | OS | OS* | |
Source only | 68.5 | 71.8 | 59.6 | 63.2 | 54.8 | 58.1 | 88.2 | 90.5 | 49.8 | 53.2 | 92.6 | 92.9 | 68.9 | 71.6 |
DAN [21] | 78.7 | 79.5 | 73.4 | 74.2 | 59.6 | 62.1 | 87.8 | 88.6 | 62.2 | 64.5 | 96.5 | 96.8 | 76.3 | 77.6 |
DANN [8] | 79.7 | 80.5 | 77.2 | 79.3 | 55.2 | 56.1 | 90.7 | 91.3 | 65.7 | 67.0 | 97.9 | 98.3 | 78.0 | 78.8 |
ATI [4] | 79.8 | 79.2 | 77.6 | 76.5 | 71.3 | 70.0 | 93.5 | 93.2 | 76.7 | 76.5 | 98.3 | 99.2 | 82.9 | 82.4 |
OSBP [29] | 74.7 | 76.4 | 72.6 | 73.1 | 61.7 | 63.3 | 93.3 | 94.8 | 79.8 | 82.5 | 95.7 | 96.5 | 79.6 | 81.1 |
D-FRODA [2] | 87.4 | / | 78.1 | / | 73.6 | / | 94.4 | / | 77.1 | / | 98.5 | / | 84.9 | / |
KASE [18] | 86.6 | 88.1 | 80.1 | 80.5 | 78.6 | 79.8 | 95.4 | 95.9 | 82.2 | 83.4 | 98.9 | 98.8 | 87.0 | 87.8 |
ADAGEV | 86.7 | 89.9 | 81.3 | 85.4 | 81.3 | 83.7 | 96.4 | 96.8 | 82.9 | 84.8 | 99.1 | 99.5 | 88.4 | 89.8 |
Relative Imp | 0.1% | 2.0% | 1.5% | 6.1% | 3.4% | 4.9% | 1.0% | 0.9% | 0.9% | 1.7% | 0.2% | 0.7% | 1.6% | 2.3% |
From Table 2, the CSDA methods generally have accuracy below 80% since they are unable to address the additional unknown classes during adaptation. The OSDA methods outperform the CSDA methods in this experiment with a large margin, where the previous state-of-the-art method KASE achieves 87.0% OS accuracy and 87.8% OS* accuracy, significantly outperforming other methods. ADAGEV shows 88.4% OS and 89.8% OS* accuracies, consistently beating the previous state-of-the-art method in all domain adaptation experiments for both known and unknown classes. Our relative OS improvements regarding benchmarks vary from 1.6% to 28.3% with the average being 11.5%.
4.3 VisDA Experiments
The VisDA dataset [24] is a conventional but more difficult task in domain adaptation. The source domain contains synthetic images and the target domain contains real-world images. For fair comparison, we follow the conventions to set “bicycle,” “bus,” “car,” “motorcycle,” “train,” “truck” as 6 known categories. In alphabetical order, the first 3 remaining categories “aeroplane,” “horse,” “knife” are set as source unknown categories and the last 3 remaining categories “person,” “plant,” “skateboard” are set as target unknown categories. We report accuracies after training for 50 epoches.
Method | bicycle | bus | car | motorcycle | train | truck | UNK | OS | OS* |
---|---|---|---|---|---|---|---|---|---|
Source only | 42.9 | 65.9 | 58.9 | 80.5 | 80.2 | 8.7 | 10.1 | 49.6 | 56.2 |
DANN [8] | 46.1 | 69.0 | 56.2 | 84.4 | 82.8 | 18.2 | 52.0 | 58.4 | 59.5 |
OSBP [29] | 51.3 | 70.7 | 37.3 | 87.8 | 77.3 | 23.8 | 88.1 | 62.3 | 58.0 |
ADAGEV | 56.0 | 71.2 | 59.1 | 88.4 | 83.0 | 25.0 | 88.3 | 67.3 | 63.8 |
From Table 3, the previous state-of-the-art methods DANN and OSBP consistently outperform the model trained only on source data, while ADAGEV is able to achieve the OS accuracy gains of 17.7%, 8.9%, 5.0% and OS* accuracy gains of 7.6%, 4.3%, 5.8% comparing with the benchmarks, from which the effectiveness of the proposed method is validated again even in this difficult task. Our relative OS improvements regarding benchmarks are 35.7%, 15.2% and 8.0%.
5 Conclusion
We propose an open set domain adaptation method which models the tail of the entropy distributions using EVT and utilizes an instance-level reweighting strategy to detect and reject unknown samples. Experiments show that the method achieves the new state-of-the-art performance by beating the existing benchmarks by a large margin for three conventional domain adaptation datasets.
References
- [1] Anderson, P., He, X., Buehler, C., Teney, D., Johnson, M., Gould, S., Zhang, L.: Bottom-up and top-down attention for image captioning and visual question answering. In: CVPR (2017)
- [2] Baktash, M., Faraki, M., Drummond, T., Salzmann, M.: Learning factorized representations for open-set domain adaptation. In: ICLR (2019)
- [3] Bendale, A., Boult, T.: Towards open set deep networks. In: CVPR (2016)
- [4] Busto, P.P., Gall, J.: Open set domain adaptation. In: ICCV (2017)
- [5] Cho, K., van Merriënboer, B., Gulcehre, C., Bougares, F., Schwenk, H., Bengio, Y.: Learning phrase representations using RNN encoder-decoder for statistical machine translation. In: EMNLP (2014)
- [6] Dang, S., Cao, Z., Cui, Z., Pi, Y., Liu, N.: Open set incremental learning for automatic target recognition. In: IEEE T-GRS (2019)
- [7] French, G., Mackiewicz, M., Fisher, M.H.: Self-ensembling for visual domain adaptation. In: ICLR (2017)
- [8] Ganin, Y., Lempitsky, V.S.: Unsupervised domain adaptation by backpropagation. In: ICML (2015)
- [9] Geng, C., Huang, S.J., Chen, S.: Recent advances in open set recognition: A survey. In: IEEE T-PAMI (2020)
- [10] Girshick, R., Radosavovic, I., Gkioxari, G., Dollár, P., He, K.: Detectron. https://github.com/facebookresearch/detectron (2018)
- [11] Gong, B., Shi, Y., Sha, F., Grauman, K.: Geodesic flow kernel for unsupervised domain adaptation. In: CVPR (2012)
- [12] Guo, Y., Xiao, M.: Cross language text classification via subspace co-regularized multi-view learning. In: ICML (2012)
- [13] He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2015)
- [14] Hoffman, J., Tzeng, E., Darrell, T., Saenko, K.: Simultaneous deep transfer across domains and tasks. In: ICCV (2015)
- [15] Koniusz, P., Tas, Y., Porikli, F.: Domain adaptation by mixture of alignments of second- or higher-order scatter tensors. In: CVPR (2017)
- [16] Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NIPS (2012)
- [17] Lecun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. In: Proceedings of the IEEE (1998)
- [18] Lian, Q., Li, W., Chen, L., Duan, L.: Known-class aware self-ensemble for open set domain adaptation. In: arXiv:1905.01068 (2019)
- [19] Liu, H., Cao, Z., Long, M., Wang, J., Yang, Q.: Separate to adapt: Open set domain adaptation via progressive separation. In: CVPR (2019)
- [20] Long, M., Cao, Z., Wang, J., Jordan, M.I.: Conditional adversarial domain adaptation. In: NIPS (2018)
- [21] Long, M., Zhu, H., Wang, J., Jordan, M.I.: Unsupervised domain adaptation with residual transfer networks. In: NIPS (2016)
- [22] Netzer, Y., Wang, T., Coates, A., Bissacco, A., Wu, B., Ng, A.: Reading digits in natural images with unsupervised feature learning. In: NIPS (2011)
- [23] Oza, P., Patel, V.: C2ae: Class conditioned auto-encoder for open-set recognition. In: CVPR (2019)
- [24] Peng, X., Usman, B., Kaushik, N., Hoffman, J., Wang, D., Saenko, K.: Visda: The visual domain adaptation challenge. In: arXiv:1710.06924 (2017)
- [25] Qi, F., Yang, X., Xu, C.: A unified framework for multimodal domain adaptation. In: ACM Multimedia (2018)
- [26] Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE T-PAMI (2015)
- [27] Saenko, K., Kulis, B., Fritz, M., Darrell, T.: Adapting visual category models to new domains. In: ECCV (2010)
- [28] Saito, K., Watanabe, K., Ushiku, Y., Harada, T.: Maximum classifier discrepancy for unsupervised domain adaptation. In: CVPR (2018)
- [29] Saito, K., Yamamoto, S., Ushiku, Y., Harada, T.: Open set domain adaptation by backpropagation. In: ECCV (2018)
- [30] Shen, J., Qu, Y., Zhang, W., Yu, Y.: Wasserstein distance guided representation learning for domain adaptation. In: AAAI (2017)
- [31] Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: arXiv:1409.1556 (2014)
- [32] Singh, A., Natarajan, V., Shah, M., Jiang, Y., Chen, X., Batra, D., Parikh, D., Rohrbach, M.: Towards VQA models that can read. In: CVPR (2019)
- [33] Tarvainen, A., Valpola, H.: Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In: ICLR (2017)
- [34] Tzeng, E., Hoffman, J., Saenko, K., Darrell, T.: Adversarial discriminative domain adaptation. In: CVPR (2017)
- [35] Xie, S., Girshick, R.B., Dollár, P., Tu, Z., He, K.: Aggregated residual transformations for deep neural networks. In: CVPR (2017)
- [36] Yao, T., Pan, Y., Ngo, C.W., Li, H., Mei, T.: Semi-supervised domain adaptation with subspace learning for visual recognition. In: CVPR (2015)
- [37] Zhang, H., Patel, V.M.: Sparse representation-based open set recognition. In: IEEE T-PAMI (2017)
- [38] Zhang, Y., Hare, J., Prügel-Bennett, A.: Learning to count objects in natural images for visual question answering. In: ICML (2018)