Heterogeneous Domain Adaptation with Positive and Unlabeled Data
Abstract
Heterogeneous unsupervised domain adaptation (HUDA) is the most challenging domain adaptation setting where the feature spaces of source and target domains are heterogeneous, and the target domain has only unlabeled data. Existing HUDA methods assume that both positive and negative examples are available in the source domain, which may not be satisfied in some real applications. This paper addresses a new challenging setting called positive and unlabeled heterogeneous unsupervised domain adaptation (PU-HUDA), a HUDA setting where the source domain only has positives. PU-HUDA can also be viewed as an extension of PU learning where the positive and unlabeled examples are sampled from different domains. A naive combination of existing HUDA and PU learning methods is ineffective in PU-HUDA due to the gap in label distribution between the source and target domains. To overcome this issue, we propose a novel method, predictive adversarial domain adaptation (PADA), which can predict likely positive examples from the unlabeled target data and simultaneously align the feature spaces to reduce the distribution divergence between the whole source data and the likely positive target data. PADA achieves this by a unified adversarial training framework for learning a classifier to predict positive examples and a feature transformer to transform the target feature space to that of the source. Specifically, they are both trained to fool a common discriminator that determines whether the likely positive examples are from the target or source domain. We experimentally show that PADA outperforms several baseline methods, such as the naive combination of HUDA and PU learning.
Index Terms:
heterogeneous domain adaptation, PU learning, adversarial trainingI Introduction


Applying machine learning models trained with a domain to different domains may lead to performance degradation [1]. This motivates researchers to study unsupervised domain adaptation (UDA) [2, 3], which transfers knowledge from a known domain (source) where a sufficiently large number of labeled samples are available to an unknown domain (target) where only unlabeled samples are available. The UDA setting supposes that the data distributions in the source and target domains are different, but the feature space is common. However, when the data is sampled from different domains, the feature space itself often differs in real-world problems (e.g., different attributes are observed), as we show examples later. Heterogeneous unsupervised domain adaptation (HUDA) [4] is a new UDA task that addresses different feature spaces. Existing HUDA studies [5, 6, 7, 8, 9, 10] only consider situations where both positive and negative examples in a binary classification task are available in the source domain, as illustrated in Fig. 1 (a). However, in some real-world applications, such as the example below, the source domain may have access to only positive examples.
Estimation of potential customers in different industries. Consider a business alliance of a bank and a jewelry store. The goal of them is to find potential customers for the jewelry store among the bank’s large customer base. This task can be considered a simple classification problem of predicting whether a given customer in the bank’s database will purchase jewelry or not. To address this task, they use their own customer data represented by the following features. First, they both might possess basic demographics such as age and gender. On the other hand, each industry has their own features. The bank has detailed economic conditions represented by deposit balances, and the jewelry store has past purchase histories. Hence, this task can be considered a HUDA task with the jewelry store as the source and the bank as the target (see Fig. 1). However, while the bank (target) do not have information on purchases of jewelry, i.e., only has unlabeled customers, unlike a conventional HUDA setting, the jewelry store (source) has only customers who have already purchased or visited the jewelry store, i.e., positive labeled customers.
Motivated by the above application, this paper addresses a new challenging HUDA task called positive and unlabeled heterogeneous unsupervised domain adaptation (PU-HUDA), which consists of a source domain with positive data a and target domain with unlabeled data in the different feature spaces as described in Fig. 1 (b). PU-HUDA can be viewed as a hybrid setup of positive unlabeled (PU) learning [11], which trains binary classifiers from positive and unlabeled data only, and HUDA. It appears that simply combining existing HUDA and PU learning methods can properly solve PU-HUDA, but it does not work well, as we demonstrate later in our experiments. This is because applying existing HUDA methods directly to PU-HUDA minimizes the divergence between the distributions of the whole source data with positive labels only and the whole target data with positive and negative labels on the transformed feature space, resulting in a less label-discriminative feature space on the target domain as shown in Fig. 2 (top). To solve PU-HUDA properly, we need to minimize the distribution divergence between the whole source data and the target data with only positive labels on the transformed feature space as described in Fig. 2 (bottom).

To address the above issue, we regard PU-HUDA as a task to identify likely positive examples from target data and reduce the distribution divergence between the whole source examples and the target examples likely to be positive. To solve this task, we propose a new method using adversarial training as in Generative Neural Network (GAN) [12]. Our proposed method mainly consists of three models, (1) a feature transformer to transform the target feature space to that of the source, (2) a classifier to predict likely positive examples from the unlabeled target data transformed by the feature transformer, and (3) a discriminator to determine whether the likely positive examples transformed by the feature transformer are from the target or source domain. The classifier is trained to identify the likely positive examples indistinguishable from the source examples by the discriminator. The feature transformer is trained to transform the target examples predicted as positive by the classifier into the indistinguishable examples from the source examples by the discriminator. To achieve these training, we adopt an objective function based on Kullback-Leibler (KL) divergence inspired by Predictive Adversarial Networks (PAN) [13], a recently proposed PU learning method. We call this approach predictive adversarial domain adaptation (PADA). PADA can make the positive target and the whole source data closer while keeping the negative target and the whole source data away, as shown in Fig. 2 (bottom). As a result, PADA obtains a classifier that provides good classification accuracy on the feature space transformed by a feature transformer.
Our contributions can be summarized as follows:
-
•
We define a new domain adaptation task, positive and unlabeled heterogeneous unsupervised domain adaptation (PU-HUDA). To the best of our knowledge, this paper is the first study to address PU-HUDA.
-
•
We propose a novel method using adversarial training, positive-adversarial domain adaptation (PADA). The PADA’s objective is based on KL divergence inspired by PAN [13].
-
•
We experimentally show that PADA outperforms several baseline methods to solve PU-HUDA.
II Related Works
We review previous studies related to the topic of this paper and discuss how different our work is from them.
Method | Some common samples | Some common features | Source negative examples |
---|---|---|---|
are necessary? | are necessary? | are necessary? | |
KCCA [5] | Yes | No | Yes |
HHTL [6] | Yes | No | Yes |
DSFT [7] | No | Yes | Yes |
OT-based [8] | No | Yes | Yes |
SFER [9] | No | No | Yes |
GLG [10] | No | No | Yes |
PADA(this paper) | No | Yes | No |
II-A Heterogeneous Unsupervised Domain Adaptation
The common idea of existing HUDA techniques is to learn feature transformers to transform the source and target data into a homogeneous feature space to minimize the distribution distance of the two domains on that feature space. Due to this difficulty of dealing with two heterogeneous feature spaces, most HUDA techniques require some supplementary information to bridge the two domains. For example, KCCA [5] and HHTL [6] require paired (common) samples between the two domains, which are often difficult to obtain in real cases. DSFT [7] and OT(optimal transport)-based method [8] require common features between the two domains as in this paper111While DSFT and this paper consider the situation where the source and target each have domain-specific features, OT-based method considers the situation where only the target has domain-specific features.. CL-SCL [14] and FSR [15] require semantically equivalent word pairs and meta-features, respectively. There are some studies that do not explicitly require supplementary information, such as GLG [10] and SFER [9] although they implicitly assume that the two domains have sufficiently similar features. All of the methods listed here require both positive and negative examples in the source domain. A comparison of this paper with these methods is summarized in Table I.
II-B PU learning
PU learning trains a binary classification from positive and unlabeled examples without labeled negative examples [11]. Early PU learning methods adopt the two-step technique that first identifies reliable negative examples and then conducts supervised learning [16, 17]. Some studies propose PU learning methods by considering unlabeled data as negative data having label noise [18, 19]. Another promising branch of PU learning employs the framework of cost-sensitive learning, such as uPU [20] and nnPU [21]. Recently, Hu et al. [13] proposed the state-of-the-art PU learning method, Predictive Adversarial Networks (PAN) based on the revised architecture of GAN. Besides performance, PAN has the advantage of not requiring class prior probability.
II-C Domain Adaptation in PU learning setting
As in our study, there are several recent studies that address domain adaptation in the context of PU learning. For example, Sonntag et al. [22] considers the domain adaptation scenario where the source domain has labeled data of all classes, and the target domain has unlabeled data and a few positive labeled data in the homogeneous feature space. The setting differs from ours which is more difficult, in that the target domain has positive labeled data and the source domain has both positive and negative data, and in the homogeneity of the feature space. The scenario where both target and source domains have unlabeled and positive labeled data is also treated for a link prediction task in the homogeneous setting [23] and heterogeneous setting [24], respectively. Their settings are also different from ours where the source domain has only positive labeled data and the target domain has only unlabeled data. In addition, there are studies that address open set domain adaptation (OSDA) by considering it as PU learning [25, 26]. OSDA performs domain adaptation while also rejects target classes that are not present in the source domain as unknown. In their settings, the source domain has a positive (known) labeled data and the target domain has unlabeled data including positive and negative (unknown) data. However, they consider the homogeneous feature space unlike our study. Thus far, no study has examined the PU learning task in the HUDA setting.

III Positive and Unlabeled Heterogeneous Domain Adaptation
III-A Problem Definition
We define a problem addressed in this paper, PU-HUDA. We suppose that the source domain and target domain share some common features as in DSFT [7] (although our experimental results show that our proposed method works even when the number of common features is sufficiently small). Let and be the source data matrix with data samples and target data matrix with data samples, respectively. Here, and are the data matrices in the common feature space with dimension in the source and target domain, respectively. and are the data matrices in the source and target specific feature spaces with and dimensions, respectively. We suppose that all data in the source domain is labeled as positive while all data in the target domain is unlabeled (including both positive and negative examples). We remark that this setting is different from the conventional HUDA setting where the source domain has both positive and negative labeled data. The main tasks of the conventional HUDA and PU-HUDA are the same, i.e., to predict the binary labels of the target data. These settings are visualized in Fig. 1.
The simplest baseline for solving PU-HUDA is to use only common features and . Existing PU learning methods can be directly applied to the common features. However, since existing PU learning methods cannot use domain-specific features, this baseline method provides only limited performance. In solving PU-HUDA, PU learning that can take advantage of domain specific features in the heterogeneous setting is essential.
III-B Difficulty of PU-HUDA
One naive solution to tackle PU-HUDA is to directly combine the existing HUDA and PU learning methods. In this naive combination approach, we first train feature transformers that transform and into and in a new homogeneous feature space by using HUDA methods, such as DSFT, GLG, and SFER. Then, we can apply an existing PU learning method to the unlabeled data and the positive labeled data .
However, this naive combination approach can be ineffective due to the following the reason. HUDA methods learn feature transformers that reduce the distribution divergence between and under some metric, such as MMD. However, recall that includes only positive data while includes both positive and negative data. Due to this gap, the mapping that reduces the distribution divergence between and does not necessarily generate a feature space that gives high discriminative performance on the target data, as described in Fig. 2 (top). Indeed, we experimentally confirmed that this naive combination of HUDA and PU learning did not work in Sec. V. In the next section, we propose a new method to overcome this difficulty.
IV Proposed Approach
To overcome the difficulty described in the previous section, we need to make only the positive target data close to the source data while keeping the negative target data away from the source data in the transformed feature space.
IV-A Basic Idea
Specifically, we regard PU-HUDA as a task to identify likely positive examples from the target data and reduce the distribution divergence between the whole source examples and the likely positive target examples. To solve this task, we exploit adversarial training and thus introduce a discriminator. Hence, our proposed method mainly consists of the following three models as described in Fig. 3 (Details of Fig. 3 are explained in Sec. IV-C).
(1) Feature Transformer transforms the target feature space to that of the source. The target data is transformed into the data in the source-specific feature space via . Thus, we get and in the same feature space (see Fig. 3). Here, we employed an asymmetric transformation [4] (i.e., only target data is transformed) because it is difficult to train feature transformers for each of the source and target data, which have different label distributions and feature spaces, and therefore it is easy to fall into non-optimal solutions.
(2) Classifier identifies positive examples in the transformed unlabeled target data (classifies positive and negative target examples). We represent the output confidence scores of as , where and represent the probabilities that the input samples are predicted to be positive and negative, respectively.
(3) Discriminator determines whether the likely positive examples are from the target or source domain (negative examples are predicted as target). The same notation as applies to although positive and negative mean source and target for , respectively.
Since is trained to discriminate the source and target examples, and are trained as follows:
-
•
is trained so that the source data and the transformed target data predicted by to be positive cannot be distinguished by .
-
•
is trained to identify likely positive examples from the target data transformed by so that they cannot be distinguished from the source data that are all positive by . Therefore, can classify positive and negative examples.
and provide feedback to each other. That is, if can predict the correct positive examples, can reduce the distribution divergence between the truly positive target data and the source data, and conversely if can transform the target positive data into the data close to the source data, can predict the likely positive examples more indistinguishable from the source data. Therefore, by alternately training them with , we can obtain good and , which are used for binary classification inference on the target data.
We next consider the optimization ways to train , , and .
IV-B Naive Optimization Approach
We first consider focusing on only positive target data while ignoring negative target data and reducing the distribution divergence between the target positive data and the source data in the transformed feature space.
The standard objective function used in domain adaptation methods using adversarial learning [27, 28, 29] to optimize and is the following cross-entropy loss:
(1) |
where and are data distributions of the source and target data, , and . By maximizing Eq. IV-B, is trained to recognize source and target domains, and by minimizing Eq. IV-B, is trained to fool , i.e., to reduce the divergence between the source and target data.
Since optimizing only Eq. IV-B causes the label-gap problem described in Sec.III-B, we incorporate the outputs of into the second term to focus on the positive target data and train simultaneously. That is, we optimize the following objective function to train , , and :
(2) |
By weighting the second term with , the training of and focuses on the target data that predicts are likely positive. The denominator of the second term prevents from outputting 0 for any input. Therefore, by the weighted second term, , , and are expected to be trained as described in Sec. IV-A. We call this naive optimization approach weighted adversarial domain adaptation (WADA). At first glance, WADA seems to work. However, WADA has the following two problems.
-
•
Since WADA places no restrictions on incorrectly predicting positive examples to be negative, will predict as positive only those samples that are sufficiently likely to be positive.
-
•
Since the second term in Eq. IV-B tends to be smaller for negative examples, will overfit on positive examples, failing to transform negative target data into the features that can easily classify.
IV-C Predictive Adversarial Domain Adaptation
To address the problems in WADA, we propose predictive adversarial domain adaptation (PADA). PADA treats positive and negative examples equally by using an objective function based on KL divergence for confidence scores inspired by PAN [13], thus solving the two problems of WADA. The objective function of PADA is as follows:
(3) |
where and . Each is a data sample from . denotes the opposite output of , i.e., .
The first and second terms: cross-entropy for , corresponding to the empirical version of Eq. IV-B. Optimizing these terms has the following effects on and .
-
•
can identify the source and target data to some extent. Here, all the target data is regarded as negative for to give the ability to recognize the target data.
-
•
Since is trained adversarial to , can transform the whole target data into the data close to the source data to some extent.
The third and fourth terms: KL divergence between the outputs of and . These terms help to alleviate the label-gap problem caused by using only the first and second terms and the problems in WADA. Since the fourth term only symmetrizes the gradient by the third term and improves performance [13], only the effects of the third term are discussed below.
-
•
is trained to output similar scores to those of for the transformed target data. That is, tries to give a high score of positivity to the target example that has difficulty distinguishing from source examples, i.e., the positive target example. also gives a low score of positivity to the target example that is easy to distinguish from source examples by , i.e., the negative target example. Therefore, obtains an ability to classify target positive and negative examples.
-
•
aims to discriminate source examples and positive target examples. is trained to output the opposite scores from those of by the third term. If identifies the likely positive target example, tries to predict it as target (negative). Conversely, if identifies the likely negative target example, tries to predict it as source (positive). The latter effect is not desirable but is canceled with the effect of the second term.
-
•
is trained so that outputs similar scores to those of , i.e., the likely positive (resp. negative) target examples predicted by are indistinguishable (resp. distinguishable) from the source examples by . In other words, tries to make predict the likely positive (resp. negative) target example predicted by as source (resp. target). This allows to transform the positive target data into the data close to the source data while keeping the negative target data away from the source data. The latter is an effect not obtained with WADA.
As a result, we obtain a label-discriminative feature space in the target domain and a good classifier that works on that feature space. Algorithm 1 describes the learning algorithm of PADA using stochastic gradient descent.
We finally mention the role of common features in PADA. Common features prevent PADA falls into a trivial solution where predicts all target data to be negative. That is, common features are useful for finding some examples in the target data that are difficult for to distinguish from the source data, especially in the early stage of learning. It is clear that PADA will not fall into the other trivial solution where predicts all target data to be positive because is learned with the target data as negative by the second term of Eq. IV-C. See the next subsection for how to use common features more explicitly to facilitate the learning of .
IV-D Soft labeling mechanism to improve PADA
In the early stage of learning PADA, since is not fully trained, the features created by may make the whole features useless. As a result, might fail to identify likely positive examples from , which leads to the unstable learning of and . Therefore, in this section, we consider guiding the learning of by a base classifier that can be trained stably using only common features without .
First, we apply an existing PU learning method to the common features of the source data and target data, and . Let be the base classifier obtained from the existing PU learning method using and as the positive and unlabeled data, respectively. If the common features contain some useful information for classification, should attain some degree of classification accuracy. Therefore, we can use to guide the learning of . By using , the objective of PADA is modified as follows:
(4) |
where is a balancing parameter. Note that takes only common features as inputs. The second and third terms are added to the original objective of PADA. Since the third term is a symmetrization term, we discuss only the second term. In the second term, gives soft labels222Soft labels are usually used in the context of distillation [30], which transfers the knowledge of large models to small models. to the unlabeled data , and is trained using them as teacher labels. By this term, the learning of will be accelerated when and are not yet learned enough, which also leads to accelerated learning of and as a result.
Once the learning of is finished, is expected to perform better than . Therefore, it is also expected that training PADA again using this as a base classifier will yield a better classifier. With this insight, we adopt repeating this soft-labeling process until the accuracy of the finally produced classifier saturates. We note that, in the second and later rounds, can take both common and target-specific features as inputs.
V Experiments
V-A Experimental Setup
V-A1 Datasets
We reorganized three datasets and used them to evaluate our proposed methods since no public dataset was directly related to the PU-HUDA setting. Detailed information about the datasets is presented in Table II.
Dataset | Number of samples | Number of features | |||||||
Training | Testing | Domain-specific | |||||||
Target | Source | Pos | Neg | Common | Target | Source | |||
Pos | Neg | Pos | |||||||
Movielens-Netflix | 6,836 | 9,164 | 8,835 | 55,725 | 74,816 | 4 | 13 | 20 | |
20-Newsgroups | CR | 1,168 | 1,197 | 1,768 | 397 | 390 | 500 | 5,000 | 5,000 |
CS | 1,168 | 1,187 | 1,768 | 378 | 406 | ||||
CT | 1,168 | 842 | 1,768 | 391 | 278 | ||||
RS | 1,192 | 1,187 | 1,197 | 396 | 396 | ||||
RT | 1,192 | 842 | 1,197 | 412 | 266 | ||||
ST | 1,187 | 842 | 1,186 | 380 | 296 | ||||
Credit card | 0.5 | 3,763 | 3,763 | 2,873 | 753 | 753 | 4 | 13 | 13 |
0.3 | 3,763 | 8,780 | 2,873 | 732 | 1,778 | ||||
0.1 | 1,594 | 14,349 | 2,873 | 325 | 2,865 |
Movielens-Netflix [31, 32]: Movielens and Netflix are two movie datasets that contain users and their ratings for movies on a scale of [0,5]. To create a classification dataset using them, we focus on the movie genre rather than the movie title. We first calculate each user’s average rating for each genre, and the differences between those and the user’s average rating for all movies are used as features. The genres common to Movielens and Netflix correspond to common features. We select one feature from common features and label each user as positive or negative according to whether the value for the feature is positive or negative, i.e., whether the user likes the corresponding genre or not. We chose Adventure as the label because the correlation of Adventure with other features is high, and the ratios of positive and negative are approximately equal. Since the default proportion of common genres is large, we randomly chose 4 genres from the 16 common genres and used them as common features. Half of the remaining common genres were used for Movielens-specific features and half for Netflix-specific features. We assigned Movielens as the target and Netflix as the source.
20-Newsgroups [33]: a collection of about 20,000 documents belonging to four top categories: computer (C), recording (R), science (S), and talk (T), each containing four more subcategories. We create a binary classification task as in DSFT [7] by selecting two top categories and assigning them positive and negative labels. Half of the subcategories are used for source and half for target. Each document is represented by tf-idf features. We use the top 500 most frequent common words in the source and target domain as common features and the top 5000 frequent specific words in each domain as domain-specific features.
Default of credit card clients [34]: a collection of credit card records, including user demographics, history of payments, bill statements, and previous payments, with users’ default payments as binary labels. We assign male users as source and female users as target. User demographics are used as common features, which may often occur in reality. We use bill statements and half of the history of payments as source-specific features, and previous payments and half of the history of payments as target-specific features. We vary the positive ratio of the target data from 0.5 to 0.1 to see the sensitivity of PADA to it.
V-A2 Compared methods
We compare the performances of PADA and PADA with soft-labeling () to the three types of baseline methods in addition to WADA.
The first baseline is to apply existing PU learning methods to only common features. As existing PU learning methods, we use NaivePU, which treats the entire target data as negative, nnPU [21], and PAN [13].
The second type of baseline is DIST, which uses not only common features but also target-specific features333Note that DIST can not transfer the knowledge of the source-specific features to the target domain.. DIST only uses the soft labeling mechanism of , i.e., first gives soft labels to the unlabeled target data using the model trained by the first baseline and trains a classifier using all the target features and the soft labels as in usual distillation [30]. As the first baseline, we adopt PAN because of its high performance.
The last baseline is a direct combination of existing HUDA and PU learning methods described in Sec. III-B. This baseline first constructs a homogeneous feature space by existing HUDA methods and next applies PU learning methods to the source and target data on the constructed homogeneous feature space. As a PU learning method, we adopt PAN for the same reason as DIST. We use two HUDA methods as baselines, DSFT [7] and SFER [9]. The linear and non-linear DSFT are denoted by and , respectively. SFER is not used for 20-Newsgroups because SFER can only be applied to low dimensional features due to computational cost.
V-A3 Implementation details
The binary classifier trained by each method was tested on the target domain. We split each target data into training, validation, and testing datasets. Each method was tuned by the validation dataset and tested by the testing dataset. We implemented each method three times and used average test accuracy as the test metric. For PADA and WADA, we selected the best learning rate in [0.0001, 0.0005, 0.001, 0.005, 0.01, 0.05, 0.1], and in [0.0001, 0.0005, 0.001, 0.005, 0.01, 0.05, 0.1, 0.5, 1]. We repeated the soft-labeling of PADA until the convergence of validation accuracy. The batch size of PADA was set to 128. For a fair comparison, each method adopted a linear classifier and a linear discriminator, i.e., logistic regressions. We also used a linear transformation for a feature transformer.
V-B Results and Analysis
V-B1 Performance comparison
Settings | Methods | |||||||||
NaivePU | nnPU | PAN | DIST | SFER | WADA | PADA | ||||
A | 59.01 | 57.89 | 59.93 | 62.64 | 62.34 | 57.33 | 57.84 | 64.40 | 66.29 | 66.13 |
B | 58.40 | 57.55 | 59.99 | 62.50 | 57.09 | 58.55 | 60.53 | 61.49 | 61.90 | 63.14 |
C | 63.72 | 63.10 | 62.64 | 65.07 | 68.35 | 60.94 | 57.82 | 67.90 | 67.82 | 68.93 |
D | 62.93 | 63.26 | 65.49 | 66.85 | 69.17 | 58.71 | 62.99 | 65.52 | 70.26 | 69.70 |
E | 62.65 | 63.10 | 66.61 | 66.59 | 66.34 | 61.54 | 57.70 | 64.60 | 65.44 | 66.72 |
Average | 61.34 | 60.98 | 62.93 | 64.73 | 64.66 | 59.42 | 59.38 | 64.78 | 66.34 | 66.92 |
Labels | Methods | ||||||||||
Pos. | Neg. | NaivePU | nnPU | PAN | DIST | SFER | WADA | PADA | |||
C | R | 63.79 | 73.10 | 80.14 | 84.88 | 84.16 | 83.02 | - | 78.37 | 80.43 | 86.11 |
C | S | 57.82 | 59.91 | 64.71 | 64.75 | 66.07 | 69.81 | - | 57.63 | 65.48 | 66.54 |
C | T | 57.85 | 73.84 | 80.67 | 79.42 | 80.17 | 80.47 | - | 77.18 | 80.67 | 83.76 |
R | S | 51.43 | 50.80 | 50.93 | 50.13 | 55.72 | 51.35 | - | 53.75 | 55.81 | 56.73 |
R | T | 57.18 | 58.21 | 66.37 | 66.52 | 65.44 | 67.01 | - | 60.28 | 66.96 | 67.21 |
S | T | 52.76 | 54.19 | 66.47 | 67.06 | 65.78 | 66.17 | - | 56.14 | 66.91 | 71.40 |
Average | 56.80 | 61.68 | 68.21 | 68.79 | 69.57 | 69.64 | - | 63.89 | 69.38 | 71.96 |
Positive | Test | Methods | |||||||||
---|---|---|---|---|---|---|---|---|---|---|---|
ratio | metric | NaivePU | nnPU | PAN | DIST | SFER | WADA | PADA | |||
0.5 | Accuracy | 54.58 | 55.47 | 54.60 | 57.85 | 44.42 | 54.07 | 49.40 | 59.88 | 60.09 | 63.70 |
AUC | 58.03 | 58.21 | 58.76 | 61.72 | 37.75 | 55.93 | 48.94 | 65.25 | 63.02 | 68.42 | |
0.3 | Accuracy | 61.29 | 69.63 | 62.79 | 66.02 | 58.91 | 65.42 | 70.57 | 69.63 | 65.31 | 68.53 |
AUC | 58.76 | 60.09 | 61.68 | 66.04 | 49.33 | 45.88 | 58.73 | 67.79 | 65.55 | 68.09 | |
0.1 | Accuracy | 64.03 | 83.78 | 63.32 | 67.64 | 62.87 | 82.59 | 80.61 | 69.20 | 70.73 | 72.20 |
AUC | 59.46 | 58.24 | 61.24 | 66.74 | 70.52 | 68.51 | 36.85 | 63.03 | 65.77 | 61.93 |
The test accuracy results of all compared methods are reported in Table III, Table IV, and Table V. We also report AUC for Default of credit card clients because the positive ratio is changed to small. The best result among compared methods is written in bold.
Movielens-Netflix: Table III shows the results for Movielens-Netflix and we can see that PADA and outperform all the baseline methods in every setting. Specifically, they improve the first baselines such as PAN using only common features by about 4% on average and by about 6% in a particular setting, which is an effect of source-specific and target-specific features. Note that DIST also improves PAN. This is because the generalization performance is improved by using not only common features but also target-specific features when performing distillation by the soft labels given by PAN. Moreover, PADA outperforms DIST, which indicates that they successfully transfers the source-specific features to the target domain. Compared to PADA, WADA’s performance is modest, as expected. We can also observe the positive effect of the soft-labeling by over PADA. We also note that the naive HUDA and PU learning combination methods, such as , , and SFER, provide limited performances. Unlike DSFT, which does nothing for common features, SFER, which transforms all features, does not work at all. This fact is consistent with the difficulty of these methods described in Sec. III-A. See Sec. V-C for verification of the effectiveness of PADA by observing the distribution divergence between the positive (or negative) target and source data in the transformed feature space.
20-Newsgroups: The results for 20-Newsgroups are presented in Table IV. In this dataset, we can observe the positive effect of DSFT. In the setting C-S, for example, improves PAN by 5%. Overall, however, its improvement over PAN is modest. Although the improvement of PADA over PAN is also modest, the soft-labeling of is considerably effective in this dataset, outperforming the baseline methods in almost all settings. We note that at most three iterations of soft labeling in were sufficient. We finally mention that WADA does not work well at all in this dataset.
Default of credit card clients: Table V shows the results for Default of credit card clients. When the positive ratio is 0.5, we can see that is the best method in terms of both accuracy and AUC. When the positive ratio is 0.3, WADA outperforms PADA although also shows almost equivalent performance to WADA. Note that the accuracy of SFER is high, its AUC is low, which means it predicts most of the data as negative. In the case of 0.1 positive ratio, PADA and also show the improvements over most of the baselines. However, is better than them. As a result, DSFT may be effective in particular datasets and settings, but its effective range is very limited.
V-B2 Analysis on the number of common features

Using the Movielens-Netflix dataset, we further investigate the effect of the number of common features on our proposed methods and several baseline methods, including PAN, DIST, DSFT, and SFER. Fig. 4 shows the test accuracy when varying the number of selected common features from 2 to 16 (all genres common to Movielens and Netflix). We can see that the performance of each method degrades as the number of common features decreases, but shows the least degradation. Conversely, the greater the number of common features, the smaller the performance difference between all methods except SFER. This is because as the number of common features increases, the number of target-specific features decreases, i.e., the target-specific features lose their usefulness. SFER, a naive combination of HUDA and PU learning is not valid regardless of the number of common features, which means that the homogeneous feature spaces obtained by it lose or inhibit the usefulness of common features.
V-C Adaptation analysis of PADA
In this section, we experimentally confirm that PADA transforms the positive target data into data close to the source data while keeping the negative target data away from the source data using the Movielens-Netflix dataset.
We compare PADA with three baselines: (1) Common, which uses only the raw common features of the source and target data, (2) DSFT, a naive combination of HUDA and PU learning methods, and (3) WADA. To see the distribution distance between the source data and the positive (resp. negative) target data in the resulting feature space, we define a metric (resp. ), the test accuracy of discrimination by a test discriminator that discriminates the source data and the positive (resp. negative) target data. The test discriminator is trained on the resulting feature space after finishing the learning of each method or only on the common feature space for Common. Note that the test discriminator is different from the discriminator trained for PADA and WADA. The true labels of the target data are used to train the test discriminator. If (resp. ) is close to 0.5, the distribution divergence between the source data and the positive (resp. negative) target data is considered small. Therefore, small and large are desirable in PU-HUDA.
The results are presented in Table VI. As expected, we can see that of PADA is smaller than those of the other methods, while of PADA is larger than it. This means that the positive target data are transformed into data close to the source data while keeping the distance between the negative target data and the source data. of Common is smaller than that of PADA, but of Common is also small. As a result, the accuracy of the existing PU learning methods using only common features is low. Even for DSFT, the difference between and is small, which is consistent with the problem of the naive combination approaches described in Sec. III-B. We also note that although the difference between and of WADA is large, its is also large. Thus, if WADA is trained to make smaller, is expected to be even closer to , which is an undesirable result as described in Sec. IV-B.
VI Conclusion
We defined a new heterogeneous domain adaptation task where only positive examples are available in the source domain. To overcome the gap in label distribution between the source and target data, we proposed a novel method that integrates the feature alignment and PU learning in a unified adversarial training framework. Moreover, we experimentally showed that our proposed method outperformed several baseline methods.
Finally, it should be mentioned that this paper does not take into account the divergence between the distributions of the source common features and target common features. It is a future work to reduce the divergence between the source common features and target common features (if there is any divergence) and increase the performance of our proposed method.
Method | Accuracy of method | Accuracy of discrimination | |
---|---|---|---|
Common | 62.93 | 54.79 | 55.79 |
DSFT | 64.73 | 72.59 | 76.58 |
WADA | 64.78 | 70.73 | 77.12 |
PADA | 66.34 | 61.08 | 66.02 |
References
- [1] A. Torralba and A. A. Efros, “Unbiased look at dataset bias,” in CVPR 2011, 2011, pp. 1521–1528.
- [2] S. J. Pan and Q. Yang, “A survey on transfer learning,” IEEE Transactions on Knowledge and Data Engineering, vol. 22, no. 10, pp. 1345–1359, 2010.
- [3] M. Wang and W. Deng, “Deep visual domain adaptation: A survey,” Neurocomputing, vol. 312, pp. 135–153, 2018.
- [4] O. Day and T. M. Khoshgoftaar, “A survey on heterogeneous transfer learning,” Journal of Big Data, vol. 4, 2017.
- [5] Y.-R. Yeh, C.-H. Huang, and Y.-C. F. Wang, “Heterogeneous domain adaptation and classification by exploiting the correlation subspace,” IEEE Transactions on Image Processing, vol. 23, no. 5, pp. 2009–2018, 2014.
- [6] J. Zhou, S. Pan, I. Tsang, and Y. Yan, “Hybrid heterogeneous transfer learning through deep learning,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 28, no. 1, 2014.
- [7] P. Wei, Y. Ke, and C. K. Goh, “A general domain specific feature transfer framework for hybrid domain adaptation,” IEEE Transactions on Knowledge and Data Engineering, vol. 31, no. 8, pp. 1440–1451, 2019.
- [8] T. Aritake and H. Hino, “Domain adaptation with optimal transport for extended variable space,” in 2022 International Joint Conference on Neural Networks (IJCNN), 2022, pp. 1–9.
- [9] F. Liu, J. Lu, and G. Zhang, “Unsupervised heterogeneous domain adaptation via shared fuzzy equivalence relations,” IEEE Transactions on Fuzzy Systems, vol. 26, no. 6, pp. 3555–3568, 2018.
- [10] F. Liu, G. Zhang, and J. Lu, “Heterogeneous domain adaptation: An unsupervised approach,” IEEE Transactions on Neural Networks and Learning Systems, vol. 31, no. 12, pp. 5588–5602, 2020.
- [11] J. Bekker and J. Davis, “Learning from positive and unlabeled data: A survey,” Machine Learning, vol. 109, pp. 719–760, 2020.
- [12] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” in Advances in Neural Information Processing Systems, Z. Ghahramani, M. Welling, C. Cortes, N. Lawrence, and K. Weinberger, Eds., vol. 27. Curran Associates, Inc., 2014.
- [13] W. Hu, R. Le, B. Liu, F. Ji, J. Ma, D. Zhao, and R. Yan, “Predictive adversarial learning from positive and unlabeled data,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 9, pp. 7806–7814, 2021.
- [14] P. Prettenhofer and B. Stein, “Cross-language text classification using structural correspondence learning,” in Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics. Uppsala, Sweden: Association for Computational Linguistics, 2010, pp. 1118–1127.
- [15] K. D. Feuz and D. J. Cook, “Transfer learning across feature-rich heterogeneous feature spaces via feature-space remapping (fsr),” ACM Trans. Intell. Syst. Technol., vol. 6, no. 1, 2015.
- [16] B. Liu, W. S. Lee, P. S. Yu, and X. Li, “Partially supervised classification of text documents,” in Proceedings of the Nineteenth International Conference on Machine Learning, ser. ICML ’02. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc., 2002, p. 387–394.
- [17] H. Yu, J. Han, and K. C.-C. Chang, “Pebl: Positive example based learning for web page classification using svm,” in Proceedings of the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, ser. KDD ’02. New York, NY, USA: Association for Computing Machinery, 2002, p. 239–248.
- [18] B. Liu, Y. Dai, X. Li, W. Lee, and P. Yu, “Building text classifiers using positive and unlabeled examples,” in Third IEEE International Conference on Data Mining, 2003, pp. 179–186.
- [19] C. Gong, H. Shi, T. Liu, C. Zhang, J. Yang, and D. Tao, “Loss decomposition and centroid estimation for positive and unlabeled learning,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 3, pp. 918–932, 2021.
- [20] M. D. Plessis, G. Niu, and M. Sugiyama, “Convex formulation for learning from positive and unlabeled data,” in Proceedings of the 32nd International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, F. Bach and D. Blei, Eds., vol. 37. Lille, France: PMLR, 2015, pp. 1386–1394.
- [21] R. Kiryo, G. Niu, M. C. du Plessis, and M. Sugiyama, “Positive-unlabeled learning with non-negative risk estimator,” in Advances in Neural Information Processing Systems, I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, Eds., vol. 30. Curran Associates, Inc., 2017.
- [22] J. Sonntag, G. Behrens, and L. Schmidt-Thieme, “Positive-unlabeled domain adaptation,” in 2022 IEEE 9th International Conference on Data Science and Advanced Analytics (DSAA), 2022, pp. 1–10.
- [23] P. Mignone and G. Pio, “Positive unlabeled link prediction via transfer learning for gene network reconstruction,” in Foundations of Intelligent Systems, M. Ceci, N. Japkowicz, J. Liu, G. A. Papadopoulos, and Z. W. Raś, Eds. Cham: Springer International Publishing, 2018, pp. 13–23.
- [24] P. Mignone, G. Pio, and M. Ceci, “Distributed heterogeneous transfer learning for link prediction in the positive unlabeled setting,” in 2022 IEEE International Conference on Big Data (Big Data), 2022, pp. 5536–5541.
- [25] M. R. Loghmani, M. Vincze, and T. Tommasi, “Positive-unlabeled learning for open set domain adaptation,” Pattern Recognition Letters, vol. 136, pp. 198–204, 2020.
- [26] S. Garg, S. Balakrishnan, and Z. C. Lipton, “Domain adaptation under open set label shift,” in Advances in Neural Information Processing Systems, A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, Eds., 2022.
- [27] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, “Domain-adversarial training of neural networks,” The journal of machine learning research, vol. 17, no. 1, pp. 2096–2030, 2016.
- [28] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
- [29] H. Tang and K. Jia, “Discriminative adversarial domain adaptation,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 04, pp. 5940–5947, 2020.
- [30] G. E. Hinton, O. Vinyals, and J. Dean, “Distilling the knowledge in a neural network,” ArXiv, vol. abs/1503.02531, 2015.
- [31] “Movielens,” 2019. [Online]. Available: https://grouplens.org/datasets/movielens/
- [32] “Netflix prize data,” 2009. [Online]. Available: https://www.kaggle.com/datasets/netflix-inc/netflix-prize-data
- [33] K. Lang, “Newsweeder: Learning to filter netnews,” in Machine Learning Proceedings 1995, A. Prieditis and S. Russell, Eds. San Francisco (CA): Morgan Kaufmann, 1995, pp. 331–339.
- [34] I.-C. Yeh and C. hui Lien, “The comparisons of data mining techniques for the predictive accuracy of probability of default of credit card clients,” Expert Systems with Applications, vol. 36, no. 2, Part 1, pp. 2473–2480, 2009.