Dual-Correction Adaptation Network for
Noisy Knowledge Transfer
Abstract
Previous unsupervised domain adaptation (UDA) methods aim to promote target learning via a single-directional knowledge transfer from label-rich source domain to unlabeled target domain, while its reverse adaption from target to source has not jointly been considered yet so far. In fact, in some real teaching practice, a teacher helps students learn while also gets promotion from students to some extent, which inspires us to explore a dual-directional knowledge transfer between domains, and thus propose a Dual-Correction Adaptation Network (DualCAN) in this paper. However, due to the asymmetrical label knowledge across domains, transfer from unlabeled target to labeled source poses a more difficult challenge than the common source-to-target counterpart. First, the target pseudo-labels predicted by source commonly involve noises due to model bias, hence in the reverse adaptation, they may hurt the source performance and bring a negative target-to-source transfer. Secondly, source domain usually contains innate noises, which will inevitably aggravate the target noises, leading to noise amplification across domains. To this end, we further introduce a Noise Identification and Correction (NIC) module to correct and recycle noises in both domains. To our best knowledge, this is the first naive attempt of dual-directional adaptation for noisy UDA, and naturally applicable to noise-free UDA. A theory justification is given to state the rationality of our intuition. Empirical results confirm the effectiveness of DualCAN with remarkable performance gains over state-of-the-arts, particularly for extreme noisy tasks (e.g., + 15 on PwPr and PrRw of Office-Home).
Index Terms:
Unsupervised domain adaptation, dual adaptation, feature noise, label noise, noise correction.1 Introduction
Deep neural network has achieved remarkable success in many applications, such as image classification and semantic segmentation. However, it relies on large-scared and high-quality annotated data, which is usually difficult to collect. Unsupervised domain adaptation (UDA) [1], which aims to adopt a fully-labeled source domain to help the learning of unlabeled target domain, has attracted much attention in recent years. Most UDA methods transfer knowledge from source to target by learning domain-invariant representation across domains, mainly including discrepancy-based [2, 3] and adversarial-based methods [4, 5, 6]. Discrepancy-based methods explicitly reduce the distribution discrepancy between domains by minimizing some distance metric, such as Maximum Mean Discrepancy (MMD) [3], Correlation Alignment (CORAL) [7] and Wasserstein distance [8]. Adversarial-based methods align feature distributions across domains by adversarial training between feature generator and domain discriminator [5], or between different classifiers [2].
In real UDA tasks, the source domain usually involves noises, further giving rise to noisy UDA [9, 10]. For example, source data collected from crowd-sourced platforms or internet medias will inevitably be corrupted by noises over both features and labels. The feature noise corrupts original features and may increase the difficulty of domain alignment, while label noise worsens the expected risk of classification, thus incurs misclassification of target instances. It makes previous UDA methods easy to fail in noisy environments. Recently, some researches [9, 10, 11, 12] have been dedicated to noisy UDA learning, which can be mainly divided into two categories. One kind of methods, such as Transferable Curriculum Learning (TCL) [9] and Robust Domain Adaptation (RDA) [10], adopts small-loss criterion to separate source instances into clean and noisy parts, then transfer source knowledge to target with only clean instances detected. The other kind, including Noisy Universal Domain Adaptation (Noisy UniDA) [11] and Noise Resistible Mutual-Training (NRMT) [12], uses co-learning strategy with multiple classifiers to reduce the impact of label noise in adaptation.
Those previous UDA methods all adopt a single-directional knowledge transfer from labeled source to unlabeled target for helping the target learning. However, in some real teaching practice, a teacher helps students learn, while also gets promotion from students to some extent. Inspired by such a philosophy, the reverse adaptation from the target should intuitively be able to boost the source learning as well, especially for weak noisy sources. To our best knowledge, however, it has not been jointly considered in UDA so far. In this paper, we attempt to explore a dual-directional knowledge transfer between domains, and propose a Dual-Correction Adaptation Network (DualCAN) to achieve mutual promotion and cooperation across domains. However, it is worth noting that there is asymmetrical label knowledge between domains, since target domain contains much less label information than the source, thus the transfer from target to source poses a more difficult challenge than the source-to-target counterpart. First, the target pseudo-labels predicted by source commonly involve noises due to transfer and model bias, hence in the reverse adaptation, they may hurt the source performance, and consequently bring a negative target-to-source transfer. Secondly, the source domain usually contains innate noises in real tasks. It will inevitably aggravate the target noises, and incurs noise amplification across domains. To address those issues, we further introduce a crucial Noise Identification and Correction (NIC) module in DualCAN to correct the noises in both domains. After that, those corrected instances are further recycled in learning rather than simply discarded, in order for a full knowledge utilization, especially in high noisy environment.
In implementing DualCAN, knowledge transfer iterates back and forth between source-to-target (ST) task and target-to-source (TS) task. In ST, source knowledge is adapted to generate target pseudo-labels, and the pseudo-labels are further corrected by NIC with self-supervised knowledge. In TS, the target knowledge is transferred reversely to correct source noise and further boost source learning. With such a dual-directional knowledge transfer between domains, noises in both domains are corrected collaboratively, and performances in both domains are promoted mutually. It is analogous to the philosophy that teaching benefits both the teacher and students alike. Quite naturally, dual-directional transfer can be adopted for both noisy and noise-free UDA tasks, while its learning concept is also applicable for some other learning tasks, for example, using down-stream tasks to reversely help feature learning in self-supervised learning. The main contributions of this paper are summarized as follows:
-
•
A Dual-Correction Adaptation Network (DualCAN) is proposed for noisy UDA learning. To our best knowledge, this is the first work of dual-directional adaptation to mutually promote learning and correct noises in both domains.
-
•
The noisy instances are corrected and recycled by a Noise Identification and Correction (NIC) module, in order to prevent noise amplification aross domains, and achieve a full knowledge utilization, especially in high noisy environment.
-
•
A theory justification is given to state the rationality of DualCAN. Moreover, empirical comparisons are conducted in real-world tasks under different noisy settings, in order to confirm the effectiveness of proposal.
The rest of the paper is organized as follows, section 2 introduces the related works, section 3 gives the preliminaries, the proposed DualCAN is described in details in section 4, and the comparison results are given in section 5. Finally, section 6 is the conclusion.
2 Related Work
Domain Adaptation. In the early stage of deep domain adaptation, discrepancy-based UDA methods [3, 13] directly minimize the distribution discrepancy between domains to learn domain-invariant representations. The commonly adopted discrepancy metrics include Maximum Mean Discrepancy (MMD) [3] and Correlation Alignment (CORAL) [7], Wasserstein distance [8], and Contrastive Domain Discrepancy (CDD) [14]. Later, inspired by the practice of adversarial learning, Ganin et al. [5] propose a Domain Adversarial Neural Network (DANN) to reduce domain gap in feature-level. Since then, a series of researches have been proposed with adversarial learning. Liu et al. [15] propose a generative adversarial network to learn joint distribution of multi-domain images. Long et al. [6] conduct an adversarial adaptation model using conditional distribution information. Hoffman et al. [16] perform adversarial learning from both pixel-level and feature-level for domain adaptation. Another line of work [17, 18, 19] treats domain adaptation as semi-supervised learning, and adopt a self-training framework to boost knowledge transfer. Researches in [20, 21] have demonstrated the effectiveness of self-training in domain adaptation under reasonable assumptions. Liu et al. [22] improve standard self-training for cycle self-training with a two-head classifier sharing a feature extractor, so as to enforce pseudo-labels to generalize across domains. They commonly adopt a single-directional knowledge transfer from source to target, while the adaptation from target to source has not been considered, which should be able to reversely boost source learning as well, especially for weak noisy sources.
Noisy UDA. Ambiguous features and incorrect labels seriously influence the generalization performance of deep CNNs. Previous methods address noises mainly by designing a robust loss function [23, 24, 25], or filtering out noisy instances in the learning process [26, 27, 28, 29, 30]. When the noisy setting is introduced into UDA, the learning problem becomes much more complex, since the unreliability of target pseudo-labels will be incurred by not only domain discrepancy, but also source noises. To reduce the effect of noisy instances, one strategy is following the small-loss criterion to collect clean source data for adaptation. For example, Shu et al. [9] propose a transferable curriculum to enhance positive transfer from clean source instances, thus mitigate negative transfer by noise. Han et al. [10] improve the curriculum learning by retaining feature-corrupted data, and use a proxy distribution in adversarial network. The other uses co-learning strategy with multiple classifiers to filter out source instances with incorrect annotations. For example, Zhao et al. [12] perform mutual instance selection to select reliable instances according to peer-confidence and relationship disagreement of networks. Yu et al. [11] study universal UDA in which target domain contains unknown classes, and optimize the divergence between two classifiers to detect noisy source instances. Chen et al. [31] address noisy source-free UDA by fine-tuning the pre-trained model with both pre-generated and self-generated labels in self-supervised learning.
Different from previous methods, we focus on both feature and label noises in UDA, as well as the recycling of noisy instances to achieve full knowledge utilization in high noisy environment. Moreover, we use a two-way knowledge transfer to mutually correct noise and promote performance in both domains, which can also be applied in noise-free UDA tasks.
3 Preliminaries
3.1 Problem Statement
In UDA, both source instances and labels are given, each instance and , where is the number of source instances, is the feature dimension, and is the class number. Target data is an unlabeled set with instances. The target distribution is commonly different from source distribution in UDA, i.e., or . In real-world adaptation tasks, data collected is usually corrupted with both feature and label noises, thus we usually encounter noisy source data. Specifically, assuming a noise ratio , features of a clean instance will be corrupted by noise with a probability of , i.e, and . At the same time, clean label is corrupted according to a noise transition matrix , where denotes the probability that instances in the k-th class is incorrectly labeled as .


3.2 Overall Concept
Previous noisy UDA learning methods commonly adopt a signle-directional adaptation, i.e., transfer knowledge only from source domain to target to boost target learning. Moreover, they aim to filter out the noisy instances by some criterion, such as the small loss metric, or prediction discrepancy between different classifiers. Then only clean or trusted source data are selected for adaptation across domains, as shown in Fig. 1 (a). While in this paper, we propose dual-directional transfer and correction between domains, which is shown in Fig. 1 (b). Specifically, source knowledge is first transferred to target with a forward ST adaptation, together with noise correction for target pseudo-labels , in order to alleviate the effect of both domain discrepancy and source noises. After that, target knowledge is reversely transferred to source by a backward TS adaptation, in order to further correct both feature and label noises in source domain for better source learning. With such dual adaptation and correction repeated, noises in both domains are corrected mutually and the learning performance of both domains can be promoted gradually.
Moreover, both feature and label noises are addressed in DualCAN, since intensive feature noise may also have negative effect to learning performance. Those noisy instances are identified and further corrected, in order for a recycling of noisy data in adaptation. The details of the proposed DualCAN will be introduced in details in the next section.
4 The Proposed Methods
DualCAN is designed for UDA with both feature and label noise. It learns with dual knowledge adaptation and noise correction between domains. We first introduce the network architecture of DualCAN, and then describe the dual correction and adaptation process, as well as the NIC module for noise identification and correction in separate sub-sections, respectively.

4.1 Network Architecture of DualCAN
The network architecture of DualCAN is shown in Fig. 2. Specifically, the source and target domains share the same feature generator while different classifiers denoted as and , respectively. First, noisy source data and target data are inputted into the network, and the initial feature generator and source classifier are obtained by source classification. Then in the ST task, source knowledge is transferred to target by assigning the target pseudo-labels, and those pseudo-labels are further corrected by the NIC module to generate . The corrected target pseudo-labels, along with self-supervised knowledge over target domain, are adopted to train target classifier. Reversely in the TS task, both source features and labels are corrected by NIC, then feature generator is updated so that target classifier can correctly classify the corrected source data , in this way, transfer target knowledge to source and generate domain-invariant representation. Finally, a dual adaptation and correction framework is formed, so as to mutually transfer knowledge and correct noise across domains. Further in NIC, there are actually three steps. First, the small-loss criterion is adopted to roughly separate data into trusted and untrusted parts. Secondly, a semi-supervised clustering method is adopted to identify feature and label noises. Finally, those noises are further corrected for full knowledge transfer.
4.2 Dual Correction and Adaptation
In DualCAN, there is dual-directional knowledge transfer across domains. That is, knowledge transfer iterates between the forward source-to-target (ST) task and backward target-to-source (TS) task, so as to mutually promote learning in each domain.
The initial feature generator and source classifier are obtained by source classification. Then in the forward ST task, knowledge transfer is conducted from label-rich source domain to the label-scared target. Target pseudo-labels are first predicted by source classifier , i.e.,
(1) |
where is the k-th element of target prediction. Those pseudo-labels actually contain inaccurate labels or noises due to cross-domain discrepancy and source noises as well. As a result, are further corrected by the NIC module, generating with improved label quality. Next, the target classifier is trained with the corrected target pseudo-labels, as well as self-supervised knowledge over target domain. Specifically, for each unlabeled target instance, we conduct a consistent loss between predictions for its weakly-augmented version and a strongly-augmented version , finally, the objective for learning target classifier is written as,
(2) | ||||
where is the cross-entropy loss.
In the backward TS task, the learning goal is to transfer knowledge from corrected target domain to weak noisy source. First, both features and labels of untrusted source data are corrected by NIC, generating to make it closer to a noise-free distribution. Then, with the target classifier fixed, both feature generator and source classifier are updated by correctly classifying source instances , in order to boost the target classifier on source domain to generate domain-invariant representation. The final learning objective is
(3) | ||||
Finally, the learning algorithm of DualCAN is shown in Algorithm 1 below, and the NIC module will be described in details in the next sub-section.
Input:
Noisy source data and target data
Output:
Feature extractor ; Source classifier ; Target
classifier
4.3 Noise Identification and Correction
Due to domain discrepancy and source noise, there will be an estimation bias in the predicted target pseudo-labels, which can also be considered as noisy labels. Those target noises may hurt the source performance in the reverse adaptation. Moreover, there are actually both feature and label noises in the source domain, it will aggravate the target noises, further leading to noise amplification across domains. Thus in DualCAN, there is an extra NIC module for noise identification and correction (NIC) in each iteration. It aims to detect untrusted data, identify them as feature or label noises, and then correct them to achieve full knowledge utilization. NIC actually learns in three steps.
In the first step, both source and target data are roughly divided into trusted data and untrusted data by small-loss criterion, considering that clean instances have smaller losses than noisy instances, i.e.,
(4) |
where is the pre-defined threshold. If the loss of is smaller than , it will be treated as a clean instance, and a noise otherwise. In fact, only a small fraction of instances are selected into the trusted set in this step, in order to avoid introducing noise into a clean dataset. That is, the untrusted set still contains both clean instances and noises. For a source instance , is its given class label and it is classified by source classifier to get the classification loss, i.e., . While for target instance , is the predicted pseudo-label by source classifier, and is classified by target classifier, i.e., . Since there is no ground-truth for target instances, we actually consider that if its prediction by target classifier is inconsistent with the pseudo-label from source classifier, then the pseudo-label can be unreliable, and thus be viewed as a noisy label.
In the second step, since the trusted set contains both clean instances and noises, including feature and label noises, those different kinds of instances are further identified to reduce the negative effect. Referring to semi-supervised learning, a deep clustering method is adopted here. Specifically, the trusted data by the previous step are viewed as labeled data, while the untrusted data are treated as unlabeled data. initial clusters are first formed over trusted data in terms of their class labels, i.e., the k-th cluster consists of trusted instances that belong to the k-th class,
(5) |
The corresponding cluster center is computed as the average of instances in each cluster,
(6) |
and the cluster radius is the Euclidean distance between the farthest trusted instance and cluster center,
(7) |
Then for an untrusted instance , its cluster index is assigned by the nearest cluster, i.e.,
(8) |
For a source instance in the untrusted set it will be identified as,
(9) |
where , and denote the feature noise, label noise and clean source sets, respectively. Specifically, if is outside the radius of the nearest cluster , then has no obvious common features with the trusted data, thus it is identified as feature noise. If is within the radius but its label is different from the cluster center, then it has high feature similarity with trusted data but a wrong label, finally it is identified as label noise. The other instances not covered in the above two groups are assigned to clean source set with normal features and labels. For a target instance in untrusted set, its predicted pseudo-label will be identified as a label noise directly.
Finally, those noisy instances are corrected after identification. Specifically, feature noises increase the difficulty of representation learning and usually give the classifier wrong prediction results, while in many cases they do not affect the whole vision in human view. To utilize the useful information in feature noises, we take the class center as disturbance feature, and perform weighted disturbance on noisy features to correct instances at the feature level, i.e.,
(10) |
where the disturbance weight controls the invasion proportion of disturbance. In practice, we dynamically adjust it during training, that is, gradually reduce to zero as training continues. In the beginning, feature noises are corrected for better domain alignment, as training proceeds, those noises become weaker, which can be adopted to improve the model robustness. For label noises, the corrected label is directly specified by the nearest cluster label , i.e., .
The whole learning process of the NIC module is summarized in Algorithm 2 below.
Input:
Noisy dataset and ; Feature extractor ; Disturbance weight
Output:
Corrected dataset and
4.4 Theory Justification
In this sub-section, we give a theory justification for DualCAN, based on a noisy version of generalization bound for target domain.
Proposition 1 [10] For any hypothesis , the bound of target expected risk in noisy environments is given by
(11) |
where is the ideal combined error of , and are the true labeling functions for source and target domains, respectively, and,
(12) |
From proposition 1, The target risk in noisy UDA is bounded by three items. The first item is the empirical risk of noisy source data. In DualCAN, a reverse knowledge transfer from target to source is adopted, the aligned target data can actually be treated as augmented source data after adaptation, thus the target data structure can help correct source noises and reduce source risk. At the same time, noisy source instances are corrected and recycled in DualCAN, which can further reduce the source risk, especially under extremely noisy environment. In fact, a better source learning guarantees better target learning, and reversely, it helps correct source noises and promote source performance as well, finally a virtuous circle is formed. The second item is the distribution discrepancy across noisy source domain and target domain, and the feature noise in source domain will affect the discrepancy [10]. Considering that, feature noises in source domain are also addressed in DualCAN for better domain alignment.
Further, for any labeling functions , and , we have [32]
(13) |
Then we have
(14) | ||||
The first and second items denote the disagreement between and the source labeling function on soure and target data, respectively. The third item is the gap of source and target functions over target instances, which is actually fixed with given data. Since the source instances are labeled, the first item can be very small. As a result, we focus on the second item for smaller , where is the 0-1 loss. By reversly transfer knowledge from target to source, it is helpful to reduce the discrepancy between and over the target data.
Method | Label Corruption | Feature Corruption | Mixed Corruption | ||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
AW | WA | AD | DA | WD | DW | Avg | AW | WA | AD | DA | WD | DW | Avg | AW | WA | AD | DA | WD | DW | Avg | |
ResNet | 47.2 | 33.0 | 47.1 | 31.0 | 68.0 | 58.8 | 47.5 | 70.2 | 55.1 | 73.0 | 55.0 | 94.5 | 87.2 | 72.5 | 58.8 | 39.1 | 69.3 | 37.7 | 75.2 | 75.5 | 59.3 |
SPL | 72.6 | 50.0 | 75.3 | 38.9 | 83.3 | 64.6 | 64.1 | 75.8 | 59.7 | 75.7 | 56.7 | 93.9 | 87.8 | 74.9 | 77.3 | 57.5 | 78.4 | 47.5 | 93.4 | 83.5 | 72.9 |
MentorNet | 74.4 | 54.2 | 75.0 | 43.2 | 85.9 | 70.6 | 67.2 | 76.0 | 60.3 | 75.5 | 59.1 | 93.4 | 89.9 | 75.7 | 76.8 | 59.5 | 78.2 | 52.3 | 94.4 | 89.0 | 75.0 |
RTN | 64.6 | 56.2 | 76.1 | 49.0 | 82.7 | 71.7 | 66.7 | 81.0 | 64.6 | 81.3 | 62.3 | 95.2 | 91.0 | 79.2 | 76.7 | 56.9 | 84.1 | 56.4 | 93.0 | 86.7 | 75.6 |
DANN | 61.2 | 46.2 | 57.4 | 42.4 | 74.5 | 62.0 | 57.3 | 71.3 | 54.1 | 69.0 | 54.1 | 84.5 | 84.6 | 69.6 | 69.7 | 50.0 | 69.5 | 49.1 | 80.1 | 79.7 | 66.4 |
MDD | 74.7 | 55.1 | 76.7 | 54.3 | 89.2 | 81.6 | 71.9 | 92.9 | 66.8 | 88.0 | 70.9 | 99.8 | 96.6 | 85.8 | 88.7 | 63.1 | 81.9 | 68.5 | 94.6 | 89.3 | 81.0 |
CST | 77.1 | 57.0 | 77.9 | 58.9 | 90.3 | 82.5 | 73.9 | 93.7 | 67.1 | 88.9 | 71.2 | 99.8 | 97.0 | 86.3 | 90.8 | 64.4 | 82.9 | 68.3 | 94.8 | 90.2 | 81.9 |
TCL | 82.0 | 65.7 | 83.3 | 60.5 | 90.8 | 77.2 | 76.6 | 84.9 | 62.3 | 83.7 | 64.0 | 93.4 | 91.3 | 79.9 | 87.4 | 64.6 | 83.1 | 62.2 | 99.0 | 92.7 | 81.5 |
RDA | 89.7 | 67.2 | 92.0 | 65.5 | 96.0 | 92.7 | 83.6 | 95.1 | 68.4 | 89.4 | 72.4 | 99.8 | 97.8 | 87.2 | 93.1 | 69.5 | 92.0 | 71.5 | 99.0 | 93.1 | 86.4 |
DualCAN | 94.2 | 71.6 | 93.4 | 75.3 | 98.3 | 95.7 | 88.1 | 94.8 | 71.9 | 93.2 | 76.7 | 100.0 | 98.6 | 89.4 | 95.0 | 70.9 | 93.4 | 76.5 | 99.8 | 97.4 | 88.8 |
Method | Office-Home | Bing-Caltech | ||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ArCl | ArPr | ArRw | ClAr | ClPr | ClRw | PrAr | PrCl | PrRw | RwAr | RwCl | RwPr | Avg | BC | |
ResNet | 8.0 | 11.2 | 13.2 | 7.4 | 12.1 | 12.1 | 8.9 | 8.9 | 14.9 | 10.7 | 9.4 | 17.3 | 11.2 | 74.4 |
SPL | 11.5 | 15.2 | 16.9 | 8.8 | 15.9 | 15.3 | 17.6 | 15.2 | 22.6 | 15.5 | 11.9 | 17.2 | 15.3 | 75.3 |
MentorNet | 13.8 | 15.5 | 17.3 | 9.4 | 16.0 | 15.9 | 17.4 | 16.3 | 24.2 | 16.0 | 11.1 | 21.3 | 16.2 | 75.6 |
RTN | 8.5 | 10.9 | 14.7 | 6.9 | 11.1 | 12.8 | 10.5 | 9.8 | 15.2 | 12.0 | 9.8 | 17.5 | 11.6 | 75.8 |
DANN | 7.3 | 12.6 | 13.0 | 7.8 | 12.5 | 9.8 | 8.0 | 9.1 | 15.5 | 11.7 | 9.1 | 18.8 | 11.3 | 72.3 |
MDD | 8.8 | 11.9 | 19.7 | 13.5 | 13.9 | 10.9 | 11.4 | 10.9 | 21.5 | 12.8 | 11.5 | 17.9 | 13.7 | 78.9 |
CST | 8.6 | 12.0 | 19.3 | 11.2 | 13.4 | 13.0 | 11.8 | 11.4 | 19.7 | 13.3 | 12.0 | 17.7 | 13.6 | 79.5 |
TCL | 12.0 | 19.1 | 22.4 | 14.0 | 25.6 | 21.2 | 18.7 | 20.9 | 33.9 | 26.2 | 22.7 | 27.4 | 22.0 | 79.0 |
RDA | 17.8 | 25.2 | 29.4 | 17.7 | 28.8 | 27.0 | 21.5 | 21.3 | 34.1 | 25.5 | 22.1 | 34.3 | 25.4 | 81.7 |
DualCAN | 31.1 | 39.4 | 44.1 | 30.6 | 42.0 | 40.8 | 35.9 | 36.2 | 50.3 | 39.9 | 37.9 | 59.0 | 40.6 | 82.4 |
In summary, by dual-directional knowledge transfer between domains, as well as correction and recycling for both feature and label noises, it can be expected to achieve better performance by DualCAN.
5 Experiments
In this section, we evaluate the proposed DualCAN on three vision datasets, against the state-of-art UDA and noisy UDA methods.
5.1 Setup
Datasets. Office-31 [33] is a standard domain adaptation dataset containing 4652 images with 31 classes. It consists of three domains: Amazon (A), Webcam (W) and DSLR (D). Office-Home [34] has 15599 images with 65 classes. It contains 4 domains with large domain gaps: Artistic (Ar), Clip Art (Cl), Product (Pr) and Real-World (Rw). To introduce noise in those two clean datasets, we follow the protocol in [10] to create corrupted counterparts in 3 different ways: label corruption, feature corruption, and mixed corruption. Label corruption changes the label of each image to other random classes uniformly and equally with probability . Feature corruption corrupts image pixels with probability of by Gaussian blur and salt-and-pepper noise. Mixed corruption refers that each image is corrupted by label corruption and feature corruption with probability independently. Bing-Caltech [35] is a real noisy dataset consisting of Bing (B) and Caltech-256 (C). The Bing dataset is created by collecting images from the Bing image search engine with class labels in Caltech-256, which natually contains label and feature noises. We take Bing as the noisy source domain and Caltech-256 as the clean target domain.
Benchmark methods. We compare DualCAN with state-of-the-art methods: ResNet-50 [36], Self-Paced Learning (SPL) [37], MentorNet [28], Deep Adaptation Network (DAN) [3], Residual Transfer Network (RTN) [38], Domain Adversarial Neural Network (DANN) [5], Conditional adversarial Domain Adaptation Network (CDAN) [6], Margin Disparity Discrepancy (MDD) [39], Cycle Self-Training (CST) [22], Transferable Curriculum Learning (TCL) [9], and Robust Domain Adaptation (RDA) [10]. SPL and MentorNet are label noise processing methods, TCL and RDA are noisy domain adaptation methods, while the others are standard UDA methods.
Implementation. We use all labeled source and unlabeled target instances for training following the standard protocols in UDA [5], implement both DualCAN and comparison methods in Pytorch. We use ResNet-50 pretrained on ImageNet [40] as feature extractors, and a fully connected bottleneck layer before the classification layer. We set the sample selection pre-training epoch to 30. The small-loss threshold is usually determined by empirical or prior knowledge of noise in previous methods [9, 10]. We fix the separation ratio to set as the loss of the -th instance in most tasks. Following standard protocol in [36], we set the initial learning rate to , and decay the learning rate by 0.1 in each 30 of the 90 epochs.
Method | ArCl | ArPr | ArRw | ClAr | ClPr | ClRw | PrAr | PrCl | PrRw | RwAr | RwCl | RwPr | Avg |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ResNet | 34.9 | 50 | 58 | 37.4 | 41.9 | 46.2 | 38.5 | 31.2 | 60.4 | 53.9 | 51.2 | 59.9 | 46.1 |
DAN | 43.6 | 57 | 67.9 | 45.8 | 56.5 | 60.4 | 44 | 43.6 | 67.7 | 63.1 | 51.5 | 74.3 | 56.3 |
DANN | 45.6 | 59.3 | 70.1 | 47 | 58.5 | 60.9 | 46.1 | 43.7 | 68.5 | 63.2 | 51.8 | 76.8 | 57.6 |
CDAN | 50.7 | 70.6 | 76 | 57.6 | 70 | 70 | 57.4 | 50.9 | 77.3 | 70.9 | 56.7 | 81.6 | 65.8 |
MDD | 54.9 | 73.7 | 77.8 | 60 | 71.4 | 71.8 | 61.2 | 53.6 | 78.1 | 72.5 | 60.2 | 82.3 | 68.1 |
CST | 59 | 79.6 | 83.4 | 68.4 | 77.1 | 76.7 | 68.9 | 56.4 | 83 | 75.3 | 62.2 | 85.1 | 73 |
DualCAN | 61.7 | 79.8 | 83.1 | 69.4 | 79.8 | 77.1 | 70.5 | 57.2 | 81.7 | 77.9 | 62.5 | 83.9 | 73.7 |
Method | AW | WA | AD | DA | WD | DW | Avg |
---|---|---|---|---|---|---|---|
ResNet | 68.4 | 60.7 | 68.9 | 62.5 | 99.3 | 96.7 | 76.1 |
DAN | 80.5 | 62.8 | 78.6 | 63.6 | 99.6 | 97.1 | 80.4 |
DANN | 82 | 67.4 | 79.7 | 68.2 | 99.1 | 96.9 | 82.2 |
CDAN | 94.1 | 69.3 | 92.9 | 71 | 100 | 98.6 | 87.7 |
MDD | 94.5 | 72.2 | 93.5 | 74.6 | 100 | 98.4 | 88.9 |
CST | 89.3 | 72.7 | 93.2 | 77 | 100 | 99.2 | 88.6 |
DualCAN | 95.1 | 73.1 | 93.6 | 78.2 | 100 | 99 | 89.8 |
5.2 Results
Noisy UDA. Tables 1 show the results on Office-31 under 40 label corruption, feature corruption, and mixed corruption, respectively. In mixed corruption case, 40 means 20 feature corruption and 20 label corruption independently. From those tables, we can find that
-
•
Through addressing label noise in learning, both SPL and MentorNet obtain better performance than ResNet, however, since they directly apply source classifier over target instances without taking the domain discrepancy into account, the target performance can be further boosted.
-
•
The UDA methods, such as MDD and CST, consider the domain alignment in learning, whereas there is noise in source domain, and the target learning will be affected by the source noise, leading to negative transfer. As a result, their performance can be further boosted, and it is important to correct the source noise in domain adaptation for better knowledge transfer.
-
•
TCL and RDA consider both source noise and domain discrepancy in learning. They filter out noise instances with small-loss criterion to reduce the negative impact of noise in adaptation, thus achieve better performance than those noise processing and UDA methods.
-
•
Our DualCAN outperforms the other methods in almost all cases, including TCL and RDA, indicating the effectiveness of the proposed dual correction and adaptation.
Method | Office-31 40 Mixed Corruption | ||||||
---|---|---|---|---|---|---|---|
AW | WA | AD | DA | WD | DW | Avg | |
w/o feature correction | 94.0 | 70.7 | 92.9 | 76.0 | 99.2 | 96.6 | 88.2 |
w/o label correction | 86.1 | 63.6 | 84.2 | 68.5 | 91.6 | 88.7 | 80.5 |
w/o source correction | 85.0 | 62.5 | 83.4 | 67.6 | 90.5 | 87.5 | 79.4 |
w/o target correction | 90.2 | 67.1 | 89.2 | 73.8 | 96.2 | 93.6 | 85.0 |
DualCAN | 95.0 | 70.9 | 93.4 | 76.5 | 99.8 | 97.4 | 88.8 |
Table 2 shows the results in extremely noisy environment on Office-Home with 160 mixed corruption, including 80 label corruption and 80 feature corruption independently, as well as naive noisy environment over Bing-Caltech. From this table, we have a similar conclusion that DualCAN outperforms the other methods, since standard UDA methods usually overfit the source noise, while standard noise processing methods do not take the discrepancy across domains into account. Further, DualCAN adopts a dual correction and adaptation strategy, instead of only allowing clean samples to participate training, the noise instances are corrected and recycled in DualCAN. Finally, DualCAN achieves better performance than both TCL and RDA, demonstrating the robustness of our proposal in high noisy environment.
Noise-free UDA. Besides noisy UDA, dual-directional transfer is also applicable to noise-free UDA, thus we give the performance comparison in noise-free UDA tasks in Table 3 and 4. From those tables, it can be found that DualCAN achieves competitive performance compared with the state-of-art UDA methods. Specifically, DualCAN obtains the best performance over 5 out of the 9 adaptation tasks over Office-31, and over 9 out of the 12 tasks over Office-Home. Moreover, it achieves the best average performance on both datasets. As a result, a dual-directional knowledge transfer across domains can achieve competitive performance in noise-free UDA as well.




5.3 Analysis
Ablation Study. In order to investigate the contribution of each component in DualCAN better, we perform an ablation study on Office-31 with 40 mixed corruption by deleting some components in DualCAN, and the results are reported in Table 5. From Table 5, it can be found that abandoning either label correction or feature correction leads to performance degradation, thus it is reasonable to address both label and feature noises in learning. At the same time, label noise has a greater impact to performance, demonstrating the necessary for different treatment of label and feature noises. Furthermore, the performance of DualCAN is better than source correction or target correction alone, which indicates that the dual correction is indeed helpful for alleviating the negative effects of noisy instances, and improving the quality of pseudo-labels, thus boosts both source and target learning.

Noise Distribution. In DualCAN, a semi-supervised deep clustering is adopted to identify feature and label noises. In Fig. 3, we show the separability of feature or label noises from clean instances on Product domain of Office-Home. Specifically, the separability by traditional loss distribution in RDA is shown in Fig. 3 (a), and the separability by clustering in DualCAN is shown in Fig. 3 (b), that is, the separability by distribution of distances from instances to the nearest cluster center. In Fig. 3 (a), there are large overlaps between feature and label noises, indicating that the small-loss criterion cannot clearly separate feature and label noises. While the distance by clustering can better identify feature and label noises, as shown in Fig. 3 (b). Further, we visualize clusters of three random classes in Fig. 3 (c), in which different shapes denote instances in different classes and red triangles denote label-noisy instances that are incorrectly labeled as the other 62 categories. From this figure, it can easily be found that label noises are commonly closer to cluster centers than feature noises, and there are clear boundaries among clusters. It verifies that our proposed NIC can well identify feature and label noises for further noise correction.
Noise Levels. Fig. 4 reports the performance of individual methods in a wide range of mixed noise levels on AW task of Office-31. Specifically, the noise level is from 0.0 to 1.6, where 0.0 means the noise-free UDA scenario, and 1.6 denotes 160 mix corruption with 80 label corruption and 80 feature corruption independently. From Fig. 4, as the noise level increases, the performance decreases for all methods, especially for DANN and ResNet with no consideration of noise processing. The performance of DualCAN is more stable with the increase of noise level, and is better than those of the other methods. It is noted that when the noise level is 1.6, DualCAN performs much better than other methods, including TCL and RDA. The reason can be that DualCAN corrects and recycles noises in learning, thus can make full use of data, especially in high noisy environment. At the same time, DualCAN also achieves the best performance when the noise level is 0, which proves that our proposal is applicable in standard UDA scenarios.








Corrected Sample Quality. In order to show the quality of noise correction, we visualize the source noise ratio and the target pseudo-label error, respectively, on PrRw task of Office-Home in Fig. 5. As the training proceeds, the proportion of source noise gradually decreases, at the same time, the error of target pseudo-label decays and achieves stable in a few epochs. As a result, with the dual correction and adaptation in DualCAN, the noise in both domains can be corrected mutually, that is, the refined target knowledge can be adopted to correct the source noise, and the corrected source knowledge can further boost the target learning forming a virtuous cycle.
Feature Visualization. Fig. 6 exhibits the t-SNE embeddings [41] of the bottleneck representation by DANN, TCL, RDA and DualCAN, respectively, with 40 mixed corruption on PrRw of Office-Home. Specifically, Fig. 6 (a)-(c) show the learned target features with different colors indicating different classes, while Fig. 6 (e)-(h) display the learned target features with different colors indicating different domains. From Fig. 6 (a)-(c), the learned features from DANN are mixed up, while DualCAN distinguishes categories better than the other compared methods, including both TCL and RDA, thus it is more robust to noise. At the same time, in Fig. 6 (e)-(h), the source and target domains are better aligned by DualCAN. Moreover, DualCAN obtains clearer boundaries, better within-class compactness, and less outliers. As a result, DualCAN provides a better solution for noisy UDA.
6 Conclusion
Prevoius UDA methods commonly focus on the single-directional source-to-target knowledge transfer, while in this paper, a Dual-Correction Adaptation Network (DualCAN) is proposed with dual-directional knowledge transfer and noise correction across domains. In DualCAN, knowledge transfer iterates between source-to-target (ST) task and target-to-source (TS) task. In ST, target pseudo-labels are generated and corrected in terms of source knowledge. While in TS, target knowledge is transferred backward to correct the source noise. Moreover, both feature and label noises are addressed in DualCAN, those noisy instances are corrected and recycled by a Noise Identification and Correction (NIC) module. Empirical results indicate that DualCAN achieves remarkable performance improvement over prior state-of-the-arts.
Acknowledgments
This work is supported in part by the NSFC under Grant No.s 62076124 and 62176118.
References
- [1] S. Ben-David, J. Blitzer, K. Crammer, A. Kulesza, F. Pereira, and J. W. Vaughan, “A theory of learning from different domains,” Machine learning, vol. 79, no. 1, pp. 151–175, 2010.
- [2] K. Saito, K. Watanabe, Y. Ushiku, and T. Harada, “Maximum classifier discrepancy for unsupervised domain adaptation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 3723–3732.
- [3] M. Long, Y. Cao, J. Wang, and M. Jordan, “Learning transferable features with deep adaptation networks,” in International conference on machine learning. PMLR, 2015, pp. 97–105.
- [4] E. Tzeng, J. Hoffman, K. Saenko, and T. Darrell, “Adversarial discriminative domain adaptation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 7167–7176.
- [5] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, “Domain-adversarial training of neural networks,” The journal of machine learning research, vol. 17, no. 1, pp. 2096–2030, 2016.
- [6] M. Long, Z. Cao, J. Wang, and M. I. Jordan, “Conditional adversarial domain adaptation,” Advances in neural information processing systems, vol. 31, 2018.
- [7] B. Sun, J. Feng, and K. Saenko, “Return of frustratingly easy domain adaptation,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30, no. 1, 2016.
- [8] J. Shen, Y. Qu, W. Zhang, and Y. Yu, “Wasserstein distance guided representation learning for domain adaptation,” in Thirty-second AAAI conference on artificial intelligence, 2018.
- [9] Y. Shu, Z. Cao, M. Long, and J. Wang, “Transferable curriculum for weakly-supervised domain adaptation,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, no. 01, 2019, pp. 4951–4958.
- [10] Z. Han, X.-J. Gui, C. Cui, and Y. Yin, “Towards accurate and robust domain adaptation under noisy environments,” arXiv preprint arXiv:2004.12529, 2020.
- [11] Q. Yu, A. Hashimoto, and Y. Ushiku, “Divergence optimization for noisy universal domain adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 2515–2524.
- [12] F. Zhao, S. Liao, G.-S. Xie, J. Zhao, K. Zhang, and L. Shao, “Unsupervised domain adaptation with noise resistible mutual-training for person re-identification,” in European Conference on Computer Vision. Springer, 2020, pp. 526–544.
- [13] W. Zellinger, T. Grubinger, E. Lughofer, T. Natschläger, and S. Saminger-Platz, “Central moment discrepancy (cmd) for domain-invariant representation learning,” arXiv preprint arXiv:1702.08811, 2017.
- [14] G. Kang, L. Jiang, Y. Yang, and A. G. Hauptmann, “Contrastive adaptation network for unsupervised domain adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 4893–4902.
- [15] M.-Y. Liu and O. Tuzel, “Coupled generative adversarial networks,” Advances in neural information processing systems, vol. 29, 2016.
- [16] J. Hoffman, E. Tzeng, T. Park, J.-Y. Zhu, P. Isola, K. Saenko, A. Efros, and T. Darrell, “Cycada: Cycle-consistent adversarial domain adaptation,” in International conference on machine learning. PMLR, 2018, pp. 1989–1998.
- [17] G. French, M. Mackiewicz, and M. Fisher, “Self-ensembling for visual domain adaptation,” arXiv preprint arXiv:1706.05208, 2017.
- [18] D.-H. Lee et al., “Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks,” in Workshop on challenges in representation learning, ICML, vol. 3, no. 2, 2013, p. 896.
- [19] Y. Zou, Z. Yu, X. Liu, B. Kumar, and J. Wang, “Confidence regularized self-training,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 5982–5991.
- [20] A. Kumar, T. Ma, and P. Liang, “Understanding self-training for gradual domain adaptation,” in International Conference on Machine Learning. PMLR, 2020, pp. 5468–5479.
- [21] T. Cai, R. Gao, J. Lee, and Q. Lei, “A theory of label propagation for subpopulation shift,” in International Conference on Machine Learning. PMLR, 2021, pp. 1170–1182.
- [22] H. Liu, J. Wang, and M. Long, “Cycle self-training for domain adaptation,” Advances in Neural Information Processing Systems, vol. 34, 2021.
- [23] T. Miyato, S.-i. Maeda, M. Koyama, and S. Ishii, “Virtual adversarial training: a regularization method for supervised and semi-supervised learning,” IEEE transactions on pattern analysis and machine intelligence, vol. 41, no. 8, pp. 1979–1993, 2018.
- [24] G. Patrini, A. Rozza, A. Krishna Menon, R. Nock, and L. Qu, “Making deep neural networks robust to label noise: A loss correction approach,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 1944–1952.
- [25] S. Reed, H. Lee, D. Anguelov, C. Szegedy, D. Erhan, and A. Rabinovich, “Training deep neural networks on noisy labels with bootstrapping,” arXiv preprint arXiv:1412.6596, 2014.
- [26] B. Han, Q. Yao, X. Yu, G. Niu, M. Xu, W. Hu, I. Tsang, and M. Sugiyama, “Co-teaching: Robust training of deep neural networks with extremely noisy labels,” Advances in neural information processing systems, vol. 31, 2018.
- [27] J. Liu, R. Li, and C. Sun, “Co-correcting: noise-tolerant medical image classification via mutual label correction,” IEEE Transactions on Medical Imaging, vol. 40, no. 12, pp. 3580–3592, 2021.
- [28] L. Jiang, Z. Zhou, T. Leung, L.-J. Li, and L. Fei-Fei, “Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels,” in International Conference on Machine Learning. PMLR, 2018, pp. 2304–2313.
- [29] N. Natarajan, I. S. Dhillon, P. K. Ravikumar, and A. Tewari, “Learning with noisy labels,” Advances in neural information processing systems, vol. 26, 2013.
- [30] S. Sukhbaatar, J. Bruna, M. Paluri, L. Bourdev, and R. Fergus, “Training convolutional networks with noisy labels,” arXiv preprint arXiv:1406.2080, 2014.
- [31] W. Chen, L. Lin, S. Yang, D. Xie, S. Pu, Y. Zhuang, and W. Ren, “Self-supervised noisy label learning for source-free unsupervised domain adaptation,” arXiv preprint arXiv:2102.11614, 2021.
- [32] Y. Zhu, F. Zhuang, J. Wang, G. Ke, J. Chen, J. Bian, H. Xiong, and Q. He, “Deep subdomain adaptation network for image classification,” IEEE transactions on neural networks and learning systems, vol. 32, no. 4, pp. 1713–1722, 2020.
- [33] K. Saenko, B. Kulis, M. Fritz, and T. Darrell, “Adapting visual category models to new domains,” in European conference on computer vision. Springer, 2010, pp. 213–226.
- [34] H. Venkateswara, J. Eusebio, S. Chakraborty, and S. Panchanathan, “Deep hashing network for unsupervised domain adaptation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 5018–5027.
- [35] A. Bergamo and L. Torresani, “Exploiting weakly-labeled web images to improve object classification: a domain adaptation approach,” Advances in neural information processing systems, vol. 23, 2010.
- [36] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
- [37] M. Kumar, B. Packer, and D. Koller, “Self-paced learning for latent variable models,” Advances in neural information processing systems, vol. 23, 2010.
- [38] M. Long, H. Zhu, J. Wang, and M. I. Jordan, “Unsupervised domain adaptation with residual transfer networks,” Advances in neural information processing systems, vol. 29, 2016.
- [39] Y. Zhang, T. Liu, M. Long, and M. Jordan, “Bridging theory and algorithm for domain adaptation,” in International Conference on Machine Learning. PMLR, 2019, pp. 7404–7413.
- [40] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein et al., “Imagenet large scale visual recognition challenge,” International journal of computer vision, vol. 115, no. 3, pp. 211–252, 2015.
- [41] J. Donahue, Y. Jia, O. Vinyals, J. Hoffman, N. Zhang, E. Tzeng, and T. Darrell, “Decaf: A deep convolutional activation feature for generic visual recognition,” in International conference on machine learning. PMLR, 2014, pp. 647–655.
![]() |
Yunyun Wang Yunyun Wang received the Ph.D. degree in Computer Science and Technology from Nanjing University of Aeronautics and Astronautics in 2012. She is currently with the School of Computer Science and Technology in Nanjing University of Posts and Telecommunications. Her current research interests include pattern recognition, machine learning and neural computing. |
![]() |
Weiwen Zheng Weiwen Zheng is currently working towards the MS degree in computer science at Nanjing University of Posts and Telecommunications, Nanjing, China. Her research interests include deep learning and transfer learning. |
![]() |
Songcan Chen Song-Can Chen received his B.S. degree in mathematics from Hangzhou University (now merged into Zhejiang University), Hangzhou, in 1983. In 1985, he completed his M.S. degree in computer applications at Shanghai Jiao Tong University and then worked at Nanjing University of Aeronautics and Astronautics (NUAA), Nanjing, in January 1986. There he received his Ph.D. degree in communication and information systems in 1997. Since 1998, as a full-time professor, he has been with the College of Computer Science and Technology at NUAA. His research interests include pattern recognition, machine learning, and neural computing. He is also an IAPR Fellow. |