Unified Optimal Transport Framework for Universal Domain Adaptation
Abstract
Universal Domain Adaptation (UniDA) aims to transfer knowledge from a source domain to a target domain without any constraints on label sets. Since both domains may hold private classes, identifying target common samples for domain alignment is an essential issue in UniDA. Most existing methods require manually specified or hand-tuned threshold values to detect common samples thus they are hard to extend to more realistic UniDA because of the diverse ratios of common classes. Moreover, they cannot recognize different categories among target-private samples as these private samples are treated as a whole. In this paper, we propose to use Optimal Transport (OT) to handle these issues under a unified framework, namely UniOT. First, an OT-based partial alignment with adaptive filling is designed to detect common classes without any predefined threshold values for realistic UniDA. It can automatically discover the intrinsic difference between common and private classes based on the statistical information of the assignment matrix obtained from OT. Second, we propose an OT-based target representation learning that encourages both global discrimination and local consistency of samples to avoid the over-reliance on the source. Notably, UniOT is the first method with the capability to automatically discover and recognize private categories in the target domain for UniDA. Accordingly, we introduce a new metric -score to evaluate the performance in terms of both accuracy of common samples and clustering performance of private ones. Extensive experiments clearly demonstrate the advantages of UniOT over a wide range of state-of-the-art methods in UniDA.
1 Introduction
Deep neural networks have boosted performance in extensive computer vision tasks but still struggle to generalize well in cross-domain tasks that source and target domain data are drawn from the different data distributions. Unsupervised Domain Adaptation (UDA) [29] aims to transfer knowledge from fully labeled source to unlabeled target domain by minimizing the domain gap between source and target. However, existing UDA methods tackle the domain gap under a strong closed-set assumption that two domains share identical label sets, limiting their applications to real-world scenarios. Recently, Partial Domain Adaptation (PDA) [4] and Open Set Domain Adaptation (OSDA) [2] are proposed to relax the closed-set assumption, allowing the existence of private classes in the source and target domain respectively. However, all the above-mentioned settings heavily rely on prior knowledge where the common classes lie in the target domain. It may be not reasonable and realistic for UDA since the target domain is unsupervised.
To address the above problem, a generalized setting, termed as Universal Domain Adaptation (UniDA) [41], was proposed to allow both domains to own private classes but without knowing prior information, e.g. matched common classes and classes numbers in the target domain. Detecting common and target-private samples in the target domain is an essential issue in UniDA.
Existing UniDA methods [41, 18, 34] detect common and target-private samples by using some manually specified or hand-tuned threshold values. Therefore, these methods are not applicable to more realistic UniDA due to the diverse ratios of common categories as shown in Fig. 1(a). Moreover, most existing UniDA methods treat all target-private samples as a single class and cannot recognize different categories among them, as shown in Fig. 1(b). This paper is the first to emphasize that UniDA methods should have the capability to automatically discover and recognize private categories in the target domain.
Essentially, the common class detection and private class discovery problems can be both viewed as distribution transportation problems. Therefore, these issues can be formulated within the framework of Optimal Transport (OT), which is a promising optimization problem to seek an efficient solution for transporting one distribution to another. Even though many OT-based methods [12, 25, 40, 14] have been proposed for Unsupervised Domain Adaptation, most of them consider cross-domain sample-to-sample mapping under the closed-set condition and are not specialized in UniDA problem with unaligned label sets. Specifically, OT encourages a global mapping to mine domain statistics property for discovering intrinsic differences among common and target private samples. Additionally, the OT constraints can also avoid degenerate solutions in clustering representation problem [1, 6]. Inspired by this, OT can be a proper formulation with respect to inter-domain common class detection and intra-domain private class discovery.

In this paper, we propose a unified optimal transport framework, namely UniOT, to solve Universal Domain Adaptation from the perspectives of partial alignment for common class detection and target representation learning for private class discovery. With regards to common class detection, we propose an OT-based partial alignment to detect common samples in the target domain and develop an adaptive filling method to handle the diverse ratios of common categories. Without any predefined threshold values, it can automatically discover the intrinsic difference between common and private classes based on the global statistical information of the assignment matrix obtained from OT. For private class discovery, we propose an OT-based target representation learning that encourages both global discrimination and local consistency of samples to avoid the over-reliance on source supervision. In addition, UniOT has the capability to automatically discover and recognize categories in the target domain, benefiting from the representation learning with OT. We believe a competitive UniDA method should achieve high classification accuracy for the common class and also learn a discriminative representation for the target private class. However, existing methods do not provide any quantitative metric to evaluate the target representation performance for those unknown samples. For target-private class discovery purpose, we introduce a new evaluation metric, H3-score, considering not only common class accuracy but also the clustering performance of private classes.
Our contributions are summarized as follows: (1) We propose UniOT to handle two essential issues in UniDA, including common class detection and private class discovery. To our best knowledge, this is the first attempt to jointly consider common class detection and private class discovery in a unified framework by optimal transport. (2) We propose an OT-based partial alignment with adaptive filling for common class detection without any predefined threshold values. It can automatically adapt to more realistic UniDA scenarios, where the ratios of common classes are diverse as shown in Fig. 1(a). (3) We design an OT-based representation learning technique for private class discovery, considering both global discrimination of clusters and local consistency of samples. Unlike most existing methods that treat all target-private samples as a whole, our UniOT can automatically discover and recognize private categories in the target domain.
2 Related Work
Universal Domain Adaptation As a more generalized UDA setting, UniDA [41] is more challenging and realistic since the prior information of categories is unknown. UAN [41],CMU [18] and TNT [8] designed sample-level uncertainty criteria to measure domain transferability. Samples with lower uncertainty are encouraged for adversarial adaptation with higher weights. Most UniDA methods detect common samples with the sample-level criteria, which requires some manually specified and hand-tuned threshold values. Moreover, over-reliance on source supervision under category neglects discriminative representation in the target domain. DANCE [34] proposed neighborhood clustering as a self-supervised technique to learn features useful for discriminating “unknown” categories. DCC [24] enumerated cluster numbers of the target domain to obtain optimal cross-domain consensus clusters as common classes, but the consensus clusters are not robust enough due to the hard assignment of K-Means [27]. MATHS [7] detected the existence of the target private class by Hartigan’s dip test in the first stage, then trained hybrid prototypes by learning from K-Means assignment where a fixed hyper-parameter for K-Means was adopted. In this paper, we use OT to handle both common sample detection and target representation learning under a unified framework. It is worth noting that our method is adaptive for various unbalanced compositions of common and private classes, which does not require any predefined threshold. In addition, our method encourages both the global discrimination and the local consistency of samples for target representation learning.
Optimal Transport in Domain Adaptation Optimal transport (OT) was first proposed by Kantorovich [23] to obtain an efficient solution for moving one distribution of mass to another. Interior point methods can handle OT problem with computational complexity at least . Cuturi [11] first proposed to use Sinkorn’s algorithm [36] to compute an approximate transport coupling with an entropic regularization. This method is lightspeed and can handle large-scale problems efficiently. Later, DeepJDOT [12] used OT to address the domain adaptation problems to pair samples in the source and target domain. Accordingly, a new classifier is trained on the transported source data representation. Since classical OT does not allow partial displacement in transport plan, unbalanced optimal transport (UOT) is proposed to relax equality constraint by replacing the hard marginal constraints of OT with soft penalties using Kullback-Leibler (KL) divergence [17], which can also effectively be solved by generalized Sinkhorn’s algorithm [10]. JUMBOT [14] proposed a minibatch strategy coupled with unbalanced optimal transport, which can yield more robust behavior for handling minibatch UDA and PDA problem. Also, TS-POT [28] addressed the partial mapping problem in UDA and PDA with partial OT [15]. However, these methods consider a cross-domain sample-to-sample mapping under the closed-set condition or label-set prior, and are not suitable for the unknown category gap problem in UniDA. Besides, a sample-to-sample mapping neglects the global consensus of two domains and intrinsic differences among common and target private samples. In this paper, we propose an inter-domain partial alignment based on UOT for common class detection and intra-domain representation learning based on OT for private class discovery, where a sample-to-prototype mapping is designed to encourage a more global-aware assignment.
Clustering for Representation Learning Deep clustering has raised increasing attention in deep learning community due to its capability of representation learning for discriminative clusters. DeepCluster [5] proposed an end-to-end framework that iteratively obtains clustering assignment by K-Means [27] and updates feature representation by the assignment, while PICA [22] simultaneously learns both feature representation and clustering assignment. A two-step approach was proposed by SCAN [38], which firstly learns a better initial representation, then implements an end-to-end clustering learning in the next stage. These deep clustering methods assume the number of clusters is known beforehand. Self-labelling [1] and SwAV [6] formulated the clustering assignment problem into an OT problem for better feature representation learning. Notably, the number of clusters does not rely on the ground-truth categories, benefiting from the equality constraints in OT. Therefore, OT-based representation learning is more suitable for UniDA, where the number of private class is unknown. However, existing OT-based representation learning methods [1, 6] consider global discrimination of clusters but ignore local consistency of samples, limiting direct applications in UniDA. In this paper, we develop an OT-based target representation learning technique for both better global and local structures in UniDA.
3 Methodology
In Universal Domain Adaptation (UniDA), there is a labeled source domain and an unlabeled target domain . We use and to denote the label set of source and target domain respectively. Denote as the common label set shared by both domains. Let and represent label sets of source private and target private, respectively. UniDA aims to classify target samples into classes, where target-private samples are treated as unknown class uniformly.
UniDA aims to transfer knowledge from source to target under domain gap and category gap (i.e., different label sets among source and target domain), which makes it challenging to align common samples and reject private samples. Moreover, over-reliance on common knowledge extracted from the source domain neglects the target intrinsic structure. Therefore, in this paper, we propose a unified optimal transport framework to address UniDA problem from two perspectives, i.e., common class detection and private class discovery, as shown in Fig. 2.
3.1 Preliminary

Optimal Transport is a constrained optimization problem that seeks an efficient solution of transporting one distribution of mass to another. We now briefly recall the well-known optimal transport formulation.
Let denote the probability simplex, where is a -dimensional vector of ones. Given two simplex vectors and , we can write the transport polytope of and as follows:
(1) |
The transport polytope can also be interpreted as a set of all possible joint probabilities of , where and are two -dimensional random variables with marginal distribution and , respectively. Following SwAV [6], given a similarity matrix instead of distance matrix in [11], the coupling matrix (or joint probability) mapping to can be quantified by optimizing the following maximization problem:
(2) |
where and is the entropic regularization term. The optimal has been shown to be unique with the form , where and can be solved by Sinkhorn’s Algorithm [11].
Since satisfies the mass constraint (1) strictly, the objective is not suitable for partial displacement problem. In light of this, Unbalanced OT is proposed to relax the conservation of marginal constraints by allowing the system to use soft penalties [10]. The unbalanced OT is formulated as
(3) |
where is Kullback-Leibler Divergence. Then the optimization problem (3) can be solved by generalized Sinkhorn’s algorithm [10].
3.2 Inter-domain Partial Alignment for Common Class Detection
To extract common knowledge, we propose an Unbalanced OT-based Common Class Detection (CCD) method, which considers a partial mapping problem between target samples and source prototypes from the perspective of cross domain. Note that our method uses prototypes to encourage a more global-aware alignment, which discovers intrinsic differences among common and target-private samples by exploiting the statistical information of the assignment matrix obtained from the optimal transport. Since UniDA allows private classes in both source and target domains, partial alignment should be considered to avoid misalignment between target-private samples and source classes. Hence we formulate the problem mapping target features to the source prototypes with unbalanced OT objective to obtain the optimal assignment matrix , i.e.
(4) |
where refers to cosine similarity between target samples and source prototypes, and the mini-batch target -normalized features with batch-size are extracted by feature extractor , i.e. . Here we use the same prototype-based classifier in [34] to learn source prototypes. Note that the private samples will be assigned with relatively low weights in . Based on this observation, our CCD selects target samples with top confidence as common classes.
Firstly, the assignment matrix is normalized as . Then we discover statistical property from normalized and generate scores from the statistical perspectives of both target samples and source prototypes. For the -th row in , it can be seen as a vanilla prediction probability vector and we can get pseudo-label by argmax operation. And we assign target samples confidence score with the maximum value of the -th row in , i.e.
(5) |
A higher score means is relatively closer to a source prototype than any other samples and is more likely from a common class. Meanwhile, to select target common samples with top confidence, we also evaluate source prototypes confidence score with the sum of the -th column, i.e.
(6) |
Analogously, a higher score means is more likely a common source prototype, which is assigned to target samples more frequently. Then the common samples are detected by statistics mean, i.e.
(7) |
where indicates the sample is detected as common sample with top confidence, which can be assigned with pseudo label by argmax operation. We can use pseudo labels of selected target samples to compute inter-domain common class detection loss by standard cross-entropy loss, i.e.
(8) |
To ensure that statistics mean can discover the intrinsic difference between common and private classes, we need to guarantee that the input features for partial mapping should be sampled from the target domain distribution with enough sampling density. Therefore, we implement a FIFO memory queue [20] to save previous target features for filling mini-batch features, which avoids the limitation led by the lack of statistical property for target samples.
Adaptive filling for unbalanced proportion of positive and negative. Unfortunately, using statistics mean as a threshold in Eq. (7) may misclassify some common samples as private class in some extremely unbalanced cases, such as there are few private samples indeed. Essentially, the assignment weight is determined by the relative distance between samples and prototypes, and private samples with low similarity set negative examples for the others. To automatically detect target samples, we design an adaptive filling mechanism for positive and negative unbalanced proportions. Our strategy is to fill the gap of unbalanced proportion to be balanced like Fig. 2 depicted. Firstly, samples with high similarity larger than a rough boundary are defined as positive and the others are negative. Then we implement adaptive filling to be balanced when the proportion of positive or negative samples exceeds . For negative filling, we synthesize fake private feature by mixing up target feature and its farthest source prototypes evenly, i.e. .
For positive filling, we first do the CCD without adaptive filling by Eq.(7) to obtain confident features and then reuse these filtered confident features for positive filling. Note that the filling samples are randomly sampled from mix-up negatives or CCD positives and the filling size is the gap between unfilled positive and negative. After adaptive filling for balanced proportion, our CCD is adaptive for detecting confident common samples.
Adaptive update for marginal probability vector. The marginal probability vector in Objective (4) can be interpreted as a weight budget for source prototypes to be mapped with target samples. However, the source domain may own source private class, which is unreasonable to set the weight budget equally for each source prototype, i.e. . Therefore, we use the obtained in the last iteration to update with the moving average for adapting to the real source distribution , i.e.
(9) |
where is the sum of column in solved in the case.
3.3 Intra-domain Representation Learning for Private Class Discovery
To exploit target global and local structure and feature representation, we propose an OT-based Private Class Discovery (PCD) clustering technique that considers a full mapping problem between target sample features and target prototypes. Especially, our PCD avoids reliance on the number of target classes and encourages global discrimination of clusters as well as local consistency of samples.
To solve this, we pre-define learnable target prototypes and initialize them randomly, where larger can be determined by the larger number of target samples. Then we assign target features with the prototypes by solving the optimal transport optimization problem:
(10) |
where refers to cosine similarity between target samples and target prototypes, and is the concatenation of the mini-batch target features and their nearest neighbor features. The marginal constraint of uniform distribution enforces that on average each prototype is selected at least times in the batch.
The solution satisfies the constraint strictly, i.e. the sum of each row equals to . To obtain prototype assignment for each target sample, the soft pseudo-label matrix ensures the -th row is a probability vector. To encourage global discrimination of clusters, the learnable prototypes and the feature extractor are optimized by minimizing , i.e.
(11) |
where the cross-entropy loss is presented as
(12) |
To encourage local consistency of samples, we swap prediction between anchor feature and its nearest neighbor feature retrieved from the memory queue. Note that instead of swapped prediction between the different views of anchor image in SwAV [6], our swapped is based on anchor-neighbor swapping, i.e.
(13) |
Eventually, to encourage both global discrimination of clusters and local consistency of samples, our is presented as
(14) |
Similar to Sec.3.2, since OT encourages a more deliberative mapping under marginal constraint, we also need to fill input features with the previous target features in memory queue before solving the optimization problem (10).
3.4 A Unified Framework for UniDA by Optimal Transport
In UniDA, the partial alignment for common class detection and representation learning for private class discovery are based on UOT and OT models, respectively. Therefore, we can summarize the above two models in a unified framework, i.e. UniOT. Considering cross-entropy loss on source samples, Common Class Discovery loss and Private Class Discovery loss , our overall objective can be expressed as
(15) |
In the testing phase, only feature extractor and prototype-based classifier are preserved. We solve the optimization problem (4) for all the concatenated target features and adaptive filled features. Then compute in Eq. 5 for each sample. For those samples who satisfy are assigned with the nearest source class, where is the sum of the total size of target samples and adaptive filling size. Otherwise, the samples are marked as unknown.
4 Experiments
4.1 Experimental Settings
Dataset.
Our method will be validated in 4 popular datasets in Domain Adaptation. Office [33] contains 31 categories and about 4K images in 3 domains: Amazon(A), DSLR(D) and Webcam(W). Office-Home [39] contains 65 categories and about 15K images in 4 domains: Artistic images(Ar), Clip-Art images(Cl), Product images(Pr) and Real-World images(Rw). VisDA [31] is a large dataset which constains 12 categories in source domain with 15K synthetic images and target domain with 5K real-world images. DomainNet [30] is the largest domain adaptation dataset which contains 345 categories and about 0.6 million in 6 domains but we only use 3 domain: Painting (P), Real (R), and Sketch (S) like [18]. Note that we use A2W to denote that transfer task from Amazon to DSLR. We follow the dataset split in [18] to conduct experiments.
Evaluation metric.
For UniDA, the trade-off between the accuracy of common and private classes is important in evaluating performance. Thus, we use the H-score [18] to calculate the harmonic mean of the instance accuracy on common class and accuracy on the single private class . However, H-score treats all private samples as a single class. For target-private class discovery purposes, we further introduce H3-score, as the harmonic mean of accuracy on common class, accuracy on unknown class and Normalized Mutual Information (NMI) for target-private clusters, i.e.
(16) |
where NMI is a well-known measure for the quality of clustering, and NMI is obtained by K-Means with the ground truth of the number of private classes in the inference stage.
4.2 Experimental details
Our implementation of optimal transport solver is based on POT [16]. We employ pretrained Res-Net-50 [21] on ImageNet [13] as our initial backbone. The feature extractor consists of backbone and projection layers which are the same as [6] but BatchNorm layer is removed. We adopt the same optimizer details as [24]. The batch size is set to 36 for both source and target domains. For all experiments, the initial learning rate is set to for all new layers and for pretrained backbone. The total training steps are set to be 10K for all datasets. For the pre-defined number of target prototypes, a larger size of target domain indicates a larger . Therefore, we empirically set for Office, for Office-Home, for VisDA, for DomainNet. We default , , , , and for all datasets. We set the size of memory queue 2K for Office and Office-Home, 10K for VisDA and DomainNet.
All experiments are implemented on a GPU of NVIDIA TITAN V with 12GB. Each experiment takes about 2.5 hours. Our code is available at: https://github.com/changwxx/UniOT-for-UniDA.
4.3 Results
Comparison with state-of-the-arts. Tab. 1 and 2 show the H-score results for Office, Office-Home, VisDA and DomainNet. Our UniOT achieves the best performance on all benchmarks. H-score on non-UniDA methods are reported from [18], and UniOT methods are reported from original papers accordingly, except for DANCE [34] since they did not report H-score and we reproduce it with their released code for fair comparison, marked as ‡. In particular, with respect to H3-score in Tab. 3, our UniOT surpasses other methods by and in Office and Office-Home datasets respectively, which demonstrates that our UniOT achieves balanced performance for both common class detection and private class discovery.
Office | DomainNet | |||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
A2D | A2W | D2A | D2W | W2A | W2D | Avg | P2R | R2P | P2S | S2P | R2S | S2R | Avg | |
ResNet[21] | 49.78 | 47.92 | 48.48 | 54.94 | 48.96 | 55.60 | 50.94 | 30.06 | 28.34 | 26.95 | 26.95 | 26.89 | 29.74 | 28.15 |
DANN[19] | 50.18 | 48.82 | 47.69 | 52.73 | 49.33 | 54.87 | 50.60 | 31.18 | 29.33 | 27.84 | 27.84 | 27.77 | 30.84 | 29.13 |
RTN[26] | 50.18 | 50.21 | 47.65 | 54.68 | 49.28 | 55.24 | 51.21 | 32.27 | 30.29 | 28.71 | 28.71 | 28.63 | 31.90 | 30.08 |
IWAN[42] | 50.64 | 50.13 | 49.65 | 54.06 | 49.79 | 55.44 | 51.62 | 35.38 | 33.02 | 31.15 | 31.15 | 31.06 | 34.94 | 32.78 |
PADA[4] | 50.00 | 49.65 | 42.87 | 52.62 | 49.17 | 55.60 | 49.98 | 28.92 | 27.32 | 26.03 | 26.03 | 25.97 | 28.62 | 27.15 |
ATI[3] | 50.48 | 48.58 | 48.48 | 55.01 | 48.98 | 55.45 | 51.16 | 32.59 | 30.57 | 28.96 | 28.96 | 28.89 | 32.21 | 30.36 |
OSBP[35] | 51.14 | 50.23 | 49.75 | 55.53 | 50.16 | 57.20 | 52.34 | 33.60 | 33.03 | 30.55 | 30.53 | 30.61 | 33.65 | 32.00 |
UAN[41] | 59.68 | 58.61 | 60.11 | 70.62 | 60.34 | 71.42 | 63.46 | 41.85 | 43.59 | 39.06 | 38.95 | 38.73 | 43.69 | 40.98 |
CMU[18] | 68.11 | 67.33 | 71.42 | 79.32 | 72.23 | 80.42 | 73.14 | 50.78 | 52.16 | 45.12 | 44.82 | 45.64 | 50.97 | 48.25 |
DANCE‡[34] | 72.64 | 62.43 | 63.27 | 76.29 | 57.37 | 82.79 | 66.62 | - | - | - | - | - | - | - |
DCC[24] | 88.50 | 78.54 | 70.18 | 79.29 | 75.87 | 88.58 | 80.16 | 56.90 | 50.25 | 43.66 | 44.92 | 43.31 | 56.15 | 49.20 |
TNT[8] | 85.70 | 80.40 | 83.80 | 92.00 | 79.10 | 91.20 | 85.37 | - | - | - | - | - | - | - |
UniOT | 86.97 | 88.48 | 88.35 | 98.83 | 87.60 | 96.57 | 91.13 | 59.30 | 47.79 | 51.79 | 46.81 | 48.32 | 58.25 | 52.04 |
Office-Home | VisDA | |||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Ar2Cl | Ar2Pr | Ar2Rw | Cl2Ar | Cl2Pr | Cl2Rw | Pr2Ar | Pr2Cl | Pr2Rw | Rw2Ar | Rw2Cl | Rw2Pr | Avg | ||
ResNet[21] | 44.65 | 48.04 | 50.13 | 46.64 | 46.91 | 48.96 | 47.47 | 43.17 | 50.23 | 48.45 | 44.76 | 48.43 | 47.32 | 25.44 |
DANN[19] | 42.36 | 48.02 | 48.87 | 45.48 | 46.47 | 48.37 | 45.75 | 42.55 | 48.70 | 47.61 | 42.67 | 47.40 | 46.19 | 25.65 |
RTN[26] | 38.41 | 44.65 | 45.70 | 42.64 | 44.06 | 45.48 | 42.56 | 36.79 | 45.50 | 44.56 | 39.79 | 44.53 | 42.89 | 26.02 |
IWAN[42] | 40.54 | 46.96 | 47.78 | 44.97 | 45.06 | 47.59 | 45.81 | 41.43 | 47.55 | 46.29 | 42.49 | 46.54 | 45.25 | 27.64 |
PADA[4] | 34.13 | 41.89 | 44.08 | 40.56 | 41.52 | 43.96 | 37.04 | 32.64 | 44.17 | 43.06 | 35.84 | 43.35 | 40.19 | 23.05 |
ATI[3] | 39.88 | 45.77 | 46.63 | 44.13 | 44.39 | 46.63 | 44.73 | 41.20 | 46.59 | 45.05 | 41.78 | 45.45 | 44.35 | 26.34 |
OSBP[35] | 39.59 | 45.09 | 46.17 | 45.70 | 45.24 | 46.75 | 45.26 | 40.54 | 45.75 | 45.08 | 41.64 | 46.90 | 44.48 | 27.31 |
UAN[41] | 51.64 | 51.70 | 54.30 | 61.74 | 57.63 | 61.86 | 50.38 | 47.62 | 61.46 | 62.87 | 52.61 | 65.19 | 56.58 | 30.47 |
CMU[18] | 56.02 | 56.93 | 59.15 | 66.95 | 64.27 | 67.82 | 54.72 | 51.09 | 66.39 | 68.24 | 57.89 | 69.73 | 61.60 | 34.64 |
DANCE‡[34] | 26.67 | 11.27 | 18.03 | 33.17 | 12.50 | 14.33 | 41.56 | 39.92 | 33.34 | 16.31 | 27.12 | 25.86 | 25.01 | - |
DCC[24] | 57.97 | 54.05 | 58.01 | 74.64 | 70.62 | 77.52 | 64.34 | 73.60 | 74.94 | 80.96 | 75.12 | 80.38 | 70.18 | 43.02 |
TNT[8] | 61.90 | 74.60 | 80.20 | 73.50 | 71.40 | 79.60 | 74.20 | 69.50 | 82.70 | 77.30 | 70.10 | 81.20 | 74.70 | 55.30 |
UniOT | 67.27 | 80.54 | 86.03 | 73.51 | 77.33 | 84.28 | 75.54 | 63.33 | 85.99 | 77.77 | 65.37 | 81.92 | 76.57 | 57.32 |
Office | Office-Home | |||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
A2D | A2W | D2A | D2W | W2A | W2D | Avg | Ar2Cl | Ar2Pr | Ar2Rw | Cl2Ar | Cl2Pr | Cl2Rw | Pr2Ar | Pr2Cl | Pr2Rw | Rw2Ar | Rw2Cl | Rw2Pr | Avg | |
ResNet[21] | 53.90 | 51.79 | 46.81 | 59.15 | 46.54 | 61.32 | 53.25 | 41.42 | 50.88 | 49.56 | 43.55 | 46.98 | 46.62 | 45.65 | 40.38 | 50.08 | 46.57 | 41.70 | 50.84 | 46.18 |
UAN[41] | 66.15 | 64.20 | 57.90 | 72.63 | 57.93 | 75.73 | 65.76 | 48.86 | 57.19 | 58.35 | 58.80 | 61.42 | 62.80 | 51.67 | 46.11 | 63.24 | 60.69 | 49.40 | 67.62 | 57.18 |
DANCE[34] | 73.19 | 68.53 | 67.88 | 81.09 | 65.61 | 85.70 | 73.67 | 40.92 | 40.95 | 45.84 | 29.73 | 20.26 | 36.97 | 52.63 | 48.23 | 50.13 | 22.78 | 44.89 | 58.29 | 40.97 |
DCC[24] | 84.47 | 74.80 | 63.54 | 87.09 | 69.58 | 71.55 | 75.17 | 55.64 | 78.21 | 78.18 | 44.64 | 33.77 | 69.96 | 63.77 | 53.81 | 65.10 | 63.17 | 53.58 | 80.09 | 61.66 |
UniOT | 83.69 | 85.28 | 71.46 | 91.24 | 70.93 | 90.84 | 82.24 | 60.11 | 78.72 | 79.53 | 65.83 | 75.32 | 76.83 | 68.21 | 56.83 | 80.55 | 69.62 | 58.74 | 79.84 | 70.84 |
Evaluation of the effectiveness of the proposed CCD and PCD. To evaluate the contribution of , and , we train the model with different combination of each component. As shown in Tab.4, the combination of and , i.e. , brings significant contribution since target representation is crucial for distinguishing common and private samples. Especially, contributes more than , which demonstrates the benefit of global discrimination of clusters for representation learning. Besides, we also conduct experiments for CCD without adaptive filling and denote the loss as , which verifies the effectiveness of adaptive filling design.
H-score | H3-score | ||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Office | Office-Home | Office | Office-Home | ||||||||||||
A2W | D2A | Avg (6 tasks) | Ar2Pr | Cl2Rw | Avg (12 tasks) | A2W | D2A | Avg (6 tasks) | Ar2Pr | Cl2Rw | Avg (12 tasks) | ||||
✓ | 77.98 | 87.79 | 83.57 | 71.21 | 74.24 | 69.73 | 69.95 | 67.44 | 72.94 | 64.84 | 59.30 | 58.18 | |||
✓ | ✓ | 86.81 | 86.36 | 88.44 | 73.53 | 76.49 | 71.40 | 82.95 | 67.88 | 81.18 | 73.14 | 69.37 | 65.34 | ||
✓ | ✓ | 87.71 | 84.71 | 89.04 | 80.00 | 83.83 | 76.07 | 80.36 | 66.31 | 77.68 | 78.32 | 76.73 | 69.89 | ||
✓ | ✓ | 74.18 | 72.61 | 79.74 | 79.59 | 74.24 | 75.55 | 75.38 | 62.08 | 76.18 | 78.42 | 76.12 | 70.44 | ||
✓ | ✓ | ✓ | 87.84 | 89.19 | 89.86 | 75.10 | 79.14 | 72.65 | 81.25 | 70.75 | 80.49 | 75.28 | 74.02 | 68.31 | |
✓ | ✓ | ✓ | 88.48 | 88.35 | 91.13 | 80.54 | 84.28 | 76.57 | 85.28 | 71.46 | 82.24 | 78.72 | 76.83 | 70.84 |
Robustness in realistic UniDA. We compare the performance of our UniOT with DCC, DANCE and UAN under different categories split. In this analysis, we perform on A2W of Office with a fixed common label set and target-private label set , but the different size of source-private label set . We set 10 categories for , 11 categories for and vary the number of . Fig. 3(a) shows that our UniOT always outperforms other methods. Besides, UniOT is not sensitive to the variation of source-private classes while other methods are not stable enough. Additionally, we also analyze the behavior of UniOT under the different size of target-private label set . We also perform on A2W of Office with a fixed common label set for 10 categories and source-private label set for 10 categories. Fig. 3(b) shows that our UniOT still performs better, while the other methods present sensitivity to the number of target-private classes. Therefore, UniOT is more robust to the variation of target-private classes.
Feature visualization for comparison with existing UniDA methods. We use t-SNE [37] to visualize the learned target features for Rw2Pr of Office-Home. As shown in Fig. 4, the common samples are colored in black and the private samples are colored with other non-black colors by their ground-truth classes. Fig. 4(b) shows that UAN cannot recognize different categories among target-private samples since they treat all target-private samples as a single class. Especially, our UniOT discovers the global and local structure of target-private samples. Fig. 4(d) validates that UniOT learns a better target representation with global discrimination and local consistency of samples compared to DANCE shown in Fig. 4(c).






5 Conclusion
In this paper, we have proposed to use Optimal Transport to handle common class detection and private class discovery for UniDA under a unified framework, namely UniOT. More precisely, an OT-based cross-domain partial alignment was designed to detect common class with an adaptive filling strategy to handle the diverse UniDA settings. In addition, we have proposed an OT-based target representation learning for private class discovery, which encourages both global discrimination and local consistency of samples. Experimental results on four benchmark datasets have validated the superiority of our UniOT over a wide range of state-of-the-art methods. In the future, we will modify memory model for more efficient features filling in OT.
6 Acknowledgement
This work was supported by the Shanghai Sailing Program (21YF1429400, 22YF1428800), and Shanghai Frontiers Science Center of Human-centered Artificial Intelligence (ShangHAI).
References
- [1] YM Asano, C Rupprecht, and A Vedaldi. Self-labelling via simultaneous clustering and representation learning. In International Conference on Learning Representations, 2019.
- [2] Silvia Bucci, Mohammad Reza Loghmani, and Tatiana Tommasi. On the effectiveness of image rotation for open set domain adaptation. In European Conference on Computer Vision, pages 422–438. Springer, 2020.
- [3] Pau Panareda Busto, Ahsan Iqbal, and Juergen Gall. Open set domain adaptation for image and action recognition. IEEE transactions on pattern analysis and machine intelligence, 42(2):413–429, 2018.
- [4] Zhangjie Cao, Lijia Ma, Mingsheng Long, and Jianmin Wang. Partial adversarial domain adaptation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 135–150, 2018.
- [5] Mathilde Caron, Piotr Bojanowski, Armand Joulin, and Matthijs Douze. Deep clustering for unsupervised learning of visual features. In Proceedings of the European Conference on Computer Vision (ECCV), pages 132–149, 2018.
- [6] Mathilde Caron, Ishan Misra, Julien Mairal, Priya Goyal, Piotr Bojanowski, and Armand Joulin. Unsupervised learning of visual features by contrasting cluster assignments. Advances in Neural Information Processing Systems, 33:9912–9924, 2020.
- [7] Liang Chen, Qianjin Du, Yihang Lou, Jianzhong He, Tao Bai, and Minghua Deng. Mutual nearest neighbor contrast and hybrid prototype self-training for universal domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 6248–6257, 2022.
- [8] Liang Chen, Yihang Lou, Jianzhong He, Tao Bai, and Minghua Deng. Evidential neighborhood contrastive learning for universal domain adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 6258–6267, 2022.
- [9] Xinlei Chen and Kaiming He. Exploring simple siamese representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 15750–15758, 2021.
- [10] Lenaic Chizat, Gabriel Peyré, Bernhard Schmitzer, and François-Xavier Vialard. Scaling algorithms for unbalanced optimal transport problems. Mathematics of Computation, 87(314):2563–2609, 2018.
- [11] Marco Cuturi. Sinkhorn distances: Lightspeed computation of optimal transport. Advances in neural information processing systems, 26:2292–2300, 2013.
- [12] Bharath Bhushan Damodaran, Benjamin Kellenberger, Rémi Flamary, Devis Tuia, and Nicolas Courty. Deepjdot: Deep joint distribution optimal transport for unsupervised domain adaptation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 447–463, 2018.
- [13] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248–255. Ieee, 2009.
- [14] Kilian Fatras, Thibault Séjourné, Rémi Flamary, and Nicolas Courty. Unbalanced minibatch optimal transport; applications to domain adaptation. In International Conference on Machine Learning, pages 3186–3197. PMLR, 2021.
- [15] Alessio Figalli. The optimal partial transport problem. Archive for rational mechanics and analysis, 195(2):533–560, 2010.
- [16] Rémi Flamary, Nicolas Courty, Alexandre Gramfort, Mokhtar Z. Alaya, Aurélie Boisbunon, Stanislas Chambon, Laetitia Chapel, Adrien Corenflos, Kilian Fatras, Nemo Fournier, Léo Gautheron, Nathalie T.H. Gayraud, Hicham Janati, Alain Rakotomamonjy, Ievgen Redko, Antoine Rolet, Antony Schutz, Vivien Seguy, Danica J. Sutherland, Romain Tavenard, Alexander Tong, and Titouan Vayer. Pot: Python optimal transport. Journal of Machine Learning Research, 22(78):1–8, 2021.
- [17] Charlie Frogner, Chiyuan Zhang, Hossein Mobahi, Mauricio Araya, and Tomaso A Poggio. Learning with a wasserstein loss. Advances in neural information processing systems, 28, 2015.
- [18] Bo Fu, Zhangjie Cao, Mingsheng Long, and Jianmin Wang. Learning to detect open classes for universal domain adaptation. In European Conference on Computer Vision, pages 567–583. Springer, 2020.
- [19] Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky. Domain-adversarial training of neural networks. The journal of machine learning research, 17(1):2096–2030, 2016.
- [20] Kaiming He, Haoqi Fan, Yuxin Wu, Saining Xie, and Ross Girshick. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9729–9738, 2020.
- [21] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
- [22] Jiabo Huang, Shaogang Gong, and Xiatian Zhu. Deep semantic clustering by partition confidence maximisation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8849–8858, 2020.
- [23] Leonid Kantorovitch. On the translocation of masses. Management science, 5(1):1–4, 1958.
- [24] Guangrui Li, Guoliang Kang, Yi Zhu, Yunchao Wei, and Yi Yang. Domain consensus clustering for universal domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9757–9766, 2021.
- [25] Mengxue Li, Yi-Ming Zhai, You-Wei Luo, Peng-Fei Ge, and Chuan-Xian Ren. Enhanced transport distance for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13936–13944, 2020.
- [26] Mingsheng Long, Han Zhu, Jianmin Wang, and Michael I Jordan. Unsupervised domain adaptation with residual transfer networks. Advances in neural information processing systems, 29, 2016.
- [27] J MacQueen. Classification and analysis of multivariate observations. In 5th Berkeley Symp. Math. Statist. Probability, pages 281–297, 1967.
- [28] Khai Nguyen, Dang Nguyen, Tung Pham, Nhat Ho, et al. Improving mini-batch optimal transport via partial transportation. In International Conference on Machine Learning, pages 16656–16690. PMLR, 2022.
- [29] Sinno Jialin Pan and Qiang Yang. A survey on transfer learning. IEEE Transactions on knowledge and data engineering, 22(10):1345–1359, 2009.
- [30] Xingchao Peng, Qinxun Bai, Xide Xia, Zijun Huang, Kate Saenko, and Bo Wang. Moment matching for multi-source domain adaptation. In Proceedings of the IEEE/CVF international conference on computer vision, pages 1406–1415, 2019.
- [31] Xingchao Peng, Ben Usman, Neela Kaushik, Judy Hoffman, Dequan Wang, and Kate Saenko. Visda: The visual domain adaptation challenge. arXiv preprint arXiv:1710.06924, 2017.
- [32] David MW Powers. Evaluation: from precision, recall and f-measure to roc, informedness, markedness and correlation. arXiv preprint arXiv:2010.16061, 2020.
- [33] Kate Saenko, Brian Kulis, Mario Fritz, and Trevor Darrell. Adapting visual category models to new domains. In European conference on computer vision, pages 213–226. Springer, 2010.
- [34] Kuniaki Saito, Donghyun Kim, Stan Sclaroff, and Kate Saenko. Universal domain adaptation through self supervision. Advances in neural information processing systems, 33:16282–16292, 2020.
- [35] Kuniaki Saito, Shohei Yamamoto, Yoshitaka Ushiku, and Tatsuya Harada. Open set domain adaptation by backpropagation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 153–168, 2018.
- [36] Richard Sinkhorn. Diagonal equivalence to matrices with prescribed row and column sums. The American Mathematical Monthly, 74(4):402–405, 1967.
- [37] Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
- [38] Wouter Van Gansbeke, Simon Vandenhende, Stamatios Georgoulis, Marc Proesmans, and Luc Van Gool. Scan: Learning to classify images without labels. In European Conference on Computer Vision, pages 268–285. Springer, 2020.
- [39] Hemanth Venkateswara, Jose Eusebio, Shayok Chakraborty, and Sethuraman Panchanathan. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5018–5027, 2017.
- [40] Renjun Xu, Pelen Liu, Liyan Wang, Chao Chen, and Jindong Wang. Reliable weighted optimal transport for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4394–4403, 2020.
- [41] Kaichao You, Mingsheng Long, Zhangjie Cao, Jianmin Wang, and Michael I Jordan. Universal domain adaptation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 2720–2729, 2019.
- [42] Jing Zhang, Zewei Ding, Wanqing Li, and Philip Ogunbona. Importance weighted adversarial nets for partial domain adaptation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 8156–8164, 2018.
Appendix A Supplement for experimental settings
A.1 Evaluation metric
H-score was proposed in CMU [18], which emphasizes the importance of both abilities of UniDA methods. Inspired by F1-score, H-score is defined as the harmonic mean of the instance accuracy on common class and and accuracy on the “unknown” class as:
(S17) |
A.2 Dataset split.
We follow the dataset split in [18, 41] to conduct experiments. Here we present more details about the dataset split. As we state in Section 2, we denote as the common label set shared by both domains and let and represent label sets of source private and target private, respectively.
Office. UAN [41] uses the 10 classes shared by Office-31 and Caltech-256 as the common label set , then in alphabetical order, the next 10 classes are used as the , and the reset 11 classes are used as the . Office-Home. In alphabet order, UAN [41] uses the first 10 classes as , the next 5 classes as and the rest 55 classes as . VisDA. UAN [41] uses the first 6 classes as , the next 3 classes as and the rest as . DomainNet. CMU [18] uses the first 150 classes as , the next 50 classes as and the rest as .
Appendix B Additional results
H3-score on VisDA and DomainNet. Tab. S5 shows the comparison of H3-score between UniOT and the state-of-the-arts for VisDA and DomainNet. Our UniOT surpasses the other methods by a large margin, especially for VisDA. We omit H3-score for TNT [8] and DCC [24] since they did not release code or hyper-parameters for these two datasets.
H-score with standard deviation. We report error bars after running the experiments three times for UniOT. As shown in Tab. S6, we report the averaged H-score results as well as standard deviation (std) for Office and Office-Home. We observe that the std values are generally close to zero for all transfer tasks, which demonstrates the stability of our method.
VisDA | DomainNet | |||||||
---|---|---|---|---|---|---|---|---|
P2R | R2P | P2S | S2P | R2S | S2R | Avg | ||
ResNet[21] | 34.55 | 36.21 | 33.15 | 30.84 | 31.50 | 30.76 | 35.53 | 33.00 |
UAN[41] | 30.05 | 38.39 | 37.62 | 39.65 | 34.84 | 33.17 | 49.54 | 38.87 |
DANCE[34] | 6.33 | 42.45 | 42.91 | 50.03 | 45.35 | 42.50 | 44.09 | 44.56 |
UniOT | 47.23 | 46.85 | 51.75 | 47.00 | 47.98 | 58.56 | 58.56 | 51.78 |
Office | ||||||
---|---|---|---|---|---|---|
A2D | A2W | D2A | D2W | W2A | W2D | Avg |
86.971.08 | 88.480.66 | 88.350.56 | 98.830.22 | 87.600.35 | 96.570.00 | 91.130.23 |
Office-Home | ||||||
Ar2Cl | Ar2Pr | Ar2Rw | Cl2Ar | Cl2Pr | Cl2Rw | |
67.270.19 | 80.540.42 | 86.030.27 | 73.510.58 | 77.330.64 | 84.280.14 | |
Pr2Ar | Pr2Cl | Pr2Rw | Rw2Ar | Rw2Cl | Rw2Pr | Avg |
75.540.45 | 63.330.74 | 85.990.19 | 77.770.56 | 65.370.25 | 81.920.22 | 76.570.17 |
Effect of Common Class Detection (CCD). To show that our CCD can effectively detect common samples as the training progresses, we present the evolution of Recall [32] and Specificity [32] values for D2W of Office. Recall measures the fraction of common samples that are retrieved as correct common class, while specificity measures the fraction of private samples that are not retrieved. As shown in Fig. S5(a), our CCD guarantees high recall rate and low specificity rate, which verifies that our CCD can detect common samples for domain alignment and avoid misalignment for target-private samples.




Sensitivity to hyper-parameters. Fig. S5(b) shows the sensitivity of , where is the rough boundary for splitting positive and negative in adaptive filling. For the cosine similarity of two -normalized features, the similarity value is limited from to , where higher value indicates higher similarity. To rough split positive and negative, the boundary should be high enough. We can observe that the performance is not sensitive to the different .
Fig. S5(c) shows the sensitivity of moving average factor in Eq. 9 When , the performance is not good since means no adaptive update for the marginal probability vector, which cannot fit actual budget for source prototypes. When , the performance does boost and it is not sensitive to different , which demonstrates the positive effects of adaptive update for the marginal probability vector.
Fig. S5(d) shows the sensitivity of the overall loss weight factor in Eq. 15 , which demonstrates that our method is not sensitive to the different selection of varying from to .
H-score | H3-score | |||||||
---|---|---|---|---|---|---|---|---|
VisDA | P2S | R2P | S2P | VisDA | P2S | R2P | S2P | |
44.57 | 43.20 | 45.97 | 45.39 | 44.62 | 37.96 | 43.91 | 41.91 | |
53.49 | 44.31 | 45.43 | 45.71 | 44.65 | 38.49 | 44.09 | 42.25 | |
57.32 | 51.79 | 47.79 | 46.81 | 60.33 | 47.00 | 51.75 | 47.98 |
Effect of local neighbor for representation learning. To show that the local neighbor consistency does benefit the representation learning in target domain, we conduct experiments with small-batch self-supervised learning methods, such as SimSiam [9] and SwAV [6]. Such self-supervised learning methods encourage the consistency between two augmentations of one image. We conduct the experiments on VisDA and three transfer tasks on DomainNet. For SimSiam, we replace with the SimSiam loss . For SwAV, we replace with the SwAV loss . Notably, Tab. S7 shows that directly combining self-supervised methods with domain adaptation barely makes a contribution. One possible interpretation is that the self-supervised methods mainly contribute to the case without any supervision but fail to benefit under the powerful source supervision. Therefore, learning from local neighbor performs better than augmentation-based self-supervised learning methods for UniDA.
Feature visualization for the significance of Private Class Discovery. We use t-SNE [37] to visualized the both source and target features for Rw2Pr of Office-Home for our UniOT. Notably, Office-Home is a more challenging dataset where there are 50 target-private categories but only 10 common categories. Source and target common categories are printed as colourful 0-9 digits, 5 source-common categories are printed as brown capitalized letters A-E, and the left 50 target-private categories are printed as grey triangles for simplicity. As shown in Fig. S6, we can observe that our Private Class Discovery (PCD) encourages more compact representation for common classes, which can improve accuracy for common classes, such as class-2 in the figure. Moreover, learning without PCD presents disordered representation around source-private classes, which will cause acute mis-classification for target-private samples, such as class-A, class-B, class-E printed in brown. Our PCD refines this situation significantly by encouraging self-supervision for target-private samples instead of over-reliance on source supervision.



































Qualitative illustration for cross-domain mapping matrix. To better illustrate the process of mapping from target samples to source prototypes in our UOT-based Common Class Discovery, we display the mini-batch based similarity matrix and UOT coupling matrix with more detailed mapping data display in Fig.S7 and Fig.S8. To visualize source prototypes, we select the nearest samples to the source prototypes as prototype images. The visualization is conducted on A2W in Office, in which 0-9 classes are common, 10-19 are source-private and 20-30 are target-private. As shown in Fig.S8, the diagonal of UOT coupling matrix of first 10 classes are more clear than others and the others are mapped slightly, which demonstrates that our partial alignment is applicable and accords with the ground-truth labels.
Qualitative results of target cluster examples. To better illustrate the target prototypes, we show some examples of top-16 nearest target neighbors of target-common and target-private prototypes. As shown in Fig. S9, we show common mini-clusters of 2 classes per domain. And our OT-based clustering method can help learn a more fine-grained mini-cluster, such as one backpack mini-cluster (left 1) is mainly dark and the other mini-cluster (left 2) is more colorful in Fig. S9(a).
Moreover, we show some examples of target-private clusters in Fig. S10. Although target-private samples lack supervision, our Private Class Discovery can also help them learn representation in a self-supervised manner.