two stages domain invariant representation learners solve the large co-variate shift in
unsupervised domain adaptation with two dimensional data domains
Abstract
Recent developments in the unsupervised domain adaptation (UDA) enable the unsupervised machine learning (ML) prediction for target data, thus this will accelerate real world applications with ML models such as image recognition tasks in self-driving. Researchers have reported the UDA techniques are not working well under large co-variate shift problems where e.g. supervised source data consists of handwritten digits data in monotone color and unsupervised target data colored digits data from the street view. Thus there is a need for a method to resolve co-variate shift and transfer source labelling rules under this dynamics. We perform two stages domain invariant representation learning to bridge the gap between source and target with semantic intermediate data (unsupervised). The proposed method can learn domain invariant features simultaneously between source and intermediate also intermediate and target. Finally this achieves good domain invariant representation between source and target plus task discriminability owing to source labels. This induction for the gradient descent search greatly eases learning convergence in terms of classification performance for target data even when large co-variate shift. We also derive a theorem for measuring the gap between trained models and unsupervised target labelling rules, which is necessary for the free parameters optimization. Finally we demonstrate that proposing method is superiority to previous UDA methods using 4 representative ML classification datasets including 38 UDA tasks. Our experiment will be a basis for challenging UDA problems with large co-variate shift.
1 Introduction
These days UDA is attracting attentions from researchers and engineers in ML projects, since it can automatically correct difference in the marginal distributions of the training source data (supervised) and test target data (unsupervised) and learn the better model based on labeling rule from source data. Especially domain invariant representation learning methods including the domain adversarial training of neural networks (DANNs) (Ganin et al., 2017), the correlation alignment for deep domain adaptation (Deep CoRALs) (Sun & Saenko, 2016), and the deep adaptation networks (DANs) (Long et al., 2015) have achieved performance improvements in a variety of ML tasks for instance digits image recognition and the human activity recognition (HAR) with accelerometer and gyroscope (Wilson et al., 2020). There are emerging projects in a fairly business-like setting with UDA, and demonstrated a certain level of success. For instance image semantic segmentation task in self-driving where different whether conditions data undermines ML models performance due to that distribution gap between training and testing (Liu et al., 2020).
On the other hand, researchers have reported the UDA techniques are not working well under large co-variate shift. The techniques could cope with a small amount of distribution gap, e.g. between monotone color digit images and their colored version without changing background and digit itself in other means, or between simulated binary toy data and their 30 degrees rotated version (Ganin et al., 2017). In contrast to that, reported discrimination performance was not over 25 % and had very high variance using UDA with the convolutional neural networks (CNNs) backbones between monotone color digit data and colored digit data from street view (Ganin et al., 2017). Same problem was observed between binary toy data and 50-70 degrees rotated version (See Figure 2 that we explain later. We can create much better model for 30 degrees target data using same shallow neural networks backbone (Ganin et al., 2017).). Such UDA problems with large co-variate shift are often in real-world ML tasks. We limit scope of large co-variate shift problem to two dimensional data domains causing co-variate shifts. In this setting two attributes related to data are causing co-variate shifts e.g. from monotone to color and from not on the street to on the street, the details of which are dealt with in 2.2.
In this study we propose two stages domain invariant representation learning to fill this gap. Use intermediate data (unsupervised) between source and target to ensure simultaneous domain invariance between source and intermediate data and invariance between intermediate and final target data, this greatly enhances learning convergence in terms of classification performance for target. We can usually get access to intermediate unsupervised data compared to huge burden for labelling processes involving human labour and expensive measuring equipment. We also demonstrate a theorem that allows unsupervised hyper-parameters optimisation based on the reverse validation (RV) (Zhong et al., 2010). This measures the difference between the target labelling rules and the labelling rules of the model after UDA training without any access to the target supervised labels. The UDA tends to negative transfer with inappropriate free parameters (Wang et al., 2019b), therefore theoretical supported indicator is important.
Experimental results with four datasets confirmed the superiority of the proposed method to previous studies. Datasets for comparison tests with high demand for social implementations include image recognition and accelerometer based HAR, and occupancy detection with energy consumption data measured by smarter maters in general households. This paper contributes to (1) Proposition of UDA strategy as a solution to large co-variate shift, (2) Derivation of free parameters tuning indicator, enables validation for conditional distribution difference without any access for target ground truth labels, (3) Demonstration of experimental superiority after comparison tests with benchmarks using four representative datasets.
2 Preliminary
2.1 Unsupervised Domain Adaptation
Sets of data are comprised of , , (Source domain, intermediate domain, target domain respectively and we denote as the sample sizes for source, intermediate and target.), in our setting is e.g. from (UserA, Summer) when is from (UserA, Winter) and is from (UserB, Summer). Then let denote the marginal and conditional distribution of source, defined for intermediate and target similarly. They are in the homogeneous domain adaptation assumption, namely they share same sample space but different distribution (Wilson & Cook, 2020). Also this research is in co-variate shift problem, that is generally sharing conditional distribution but different marginal distribution of co-variate (Zhao et al., 2019). The objective is learning to predict ground truth labels for using three sets of data without access to target ground truth labels, corresponds to feature extractor with arbitrary neural networks parameter and task classifier with to be optimized by gradient descent.
2.2 What is two dimensional co-variate shifts
Let’s dive into what is intermediate domain , we assume data domain is not one dimensional but is two dimensional e.g. (UserA or UserB, Accelerometer Model A and Accelerometer Model B) in HAR, (UserA or UserB, Winter or Summer) in occupancy detection problem and (Monotone or Color, In the street or not) in image recognition. In this paradigm, may be (UserA, Accelerometer Model A), (UserB, Accelerometer Model A), (UserB, Accelerometer Model B) for instance. These two dimensional factors are influencing data distribution (we call this as two dimensional co-variate shifts), energy consumption data differs between users and also seasons for instance. Basically under this two dimensional co-variate shifts problem correcting difference between domains in UDA is quite difficult but more natural case closer to businesses or real world projects. Additionally we assume gathering unsupervised data is quite easy, so UDA problem with three sets is natural and there is demand for solving this problem. Two dimensional domains assumption is novel in this field, although uses of an intermediate domain to resolve large co-variate shift are explored in (Lin et al., 2021; Zhang et al., 2019; Oshima et al., 2024) as well. Their experiments showed a synthetic intermediate domain can improve UDA (Lin et al., 2021; Zhang et al., 2019), but limited to computer vision tasks technically or experimentally. Our proposition is not limited to a specific data type. We investigated whether or not each dataset follows this assumption in the Appendix E.
3 UDA when two dimensional co-variate shifts

3.1 large co-variate shift solver: two stages domain invariant learners
To begin with previous domain invariant learning abstraction, their objective functions and optimization are below (we call this as Normal). Domain invariant learning is parallel learning including the feature extractor and task classifier’s learning labelling rule based on by minimizing and the feature extractor’s learning domain invariance between by (measurement of distribution gap between source and target, we elaborate later). The constant value is coefficient of weight to adjust the balance between task classification performance and domain invariant performance. We define as the cross entropy loss with a predicted probability for source input.
(1) | |||
(2) |
This will achieve the feature extractor and task classifier’s generalization performance for target data theoretically, since the UDA goal(expectation risk for target discrimination performance) is bounded by marginal distribution difference between source and target (corresponds to ), empirical risk for source discrimination performance (corresponds to ), conditional distribution gap (under co-variate shift problem this should be small), and so fourth. Let be a hypothesis space of VC-dimension , be empirical sample sets with size drawn from source joint distribution (features and label) and target marginal distribution (features), be the domain label (binary label identifying source or target, ), be expectations for hypothesis (task classification), be the empirical expectation, be the function outputting 1 when true inside the square brackets otherwise outputting 0, and then w.p. at least (), , the following inequality holds (Zhao et al., 2019; Ben-David et al., 2006).
(3) | |||
(4) | |||
(5) |
Previous studies such as DANNs, CoRALs and DANs have been published as embodiments of equation 1 and 2. The construction of differs across methods. DANNs define it as a loss of domain classification problem, while CoRALs use the distance between covariance matrices of . DANs build it using the multiple kernel variant of maximum mean discrepancy (MK-MMD (Gretton et al., 2012)) calculated on and , and Deep Joint Distribution Optimal Transportation (DeepJDOT) define it by wasserstein distance based on the optimal transport problem (Damodaran et al., 2018). There are many variants, e.g. Convolutional deep Domain Adaptation model for Time Series data (CoDATS) is signal processing layers and multi sources version of DANNs (Wilson et al., 2020) and CoRALs variants are using different distant metrics for correlation alignment including a euclidean distance (Zhang et al., 2018a), geodesic distance (Zhang et al., 2018b), and log euclidean distance (Wang et al., 2017). For DANNs, optimizing in a adversarial way with the domain discriminator with arbitrary neural networks’ parameters , we omitted description of . Please note that is only worked for e.g. CoRALs and DANs.
We extend equation 1 to use intermediate data, as a decomposition of distribution differences between source and intermediate data and between intermediate and terminal target data, and same optimization can be used in this formula as well.
(6) | |||
(7) |
The equation 6 says if we have intermediate data we can minimize substituted by and . Concurrent training of is tantamount to (1) aquisition of domain invariance between source and intermediate domain, discriminability for intermediate domain, (2) aquition of domain invariance between intermediate domain and target, discriminability for target, we can do each quite easier compared with normal domain invariant learning between source and target with same goal due to data domain’s semantic reason, so as a whole our strategy will facilitate learning target discriminability. Based on assumption of two dimensional data domains divergence between source and target is larger than divergence between source and intermediate, or divergence between intermediate and target, we can say . This term is the optimisation target of domain invariant representation learning, but it can be inferred that the larger it is, the more likely it is to be addicted to stale solutions (e.g. local minima) when paralleled with loss minimisation for task classification. Pseudo codes for UDA with are in Algorithm 1 and Algorithm 3 (Appendix A), though of course this can be used in other domain invariant representation learning techniques in the almost same way. We implemented two stages domain invariant learners with CoRALs as , since we can say same thing as and to avoid noisy correlation alignment between intermediate and target data in the early stages of epochs. We denote as domain labels, as the cross entropy loss, as the binary cross entropy loss, as the mean squared error and as the covariance matrix. The schematic diagram for DANNs version two stages domain invariant learners is in Figure 1 (CoRALs version in Appendix A Figure 7).
Previous paper (Oshima et al., 2024) is intuitively step-by-step version of this paper (center of Figure 1), this executes DANNs learning between source and intermediate data then second time DANNs between intermediate data (pseudo labeling by that learned task classifier) and target. We call this as Step-by-step. Very limited evaluation was conducted and achieved higher classification performance compared to normal DANNs and without adaptation model using occupancy detection data(Dataset D, explain later). There are five main differences between the method of (Oshima et al., 2024) and this paper, (1) (Oshima et al., 2024) generates noise due to erroneous answers in the pseudo-labelling step ( in Figure 1), which may hinder learning convergence, whereas our end-to-end method does not, (2) The two domain invariant representation learnings have partially independent structure, and there is no guarantee that the first learning will necessarily be good for the second learning, but our method can perform the two learnings in end-to-end, which may make it easier to keep the overall balance, (3) They introduced confidence threshold technique which has large impact on learning convergence, but these engineering tricks are not easy to tune in unsupervised settings in practical use cases, (4) Comparative experiments on four sets of data show that our method has a performance advantage on most of the settings (5/8 settings) and (5) The learning algorithm encompasses the entire domain invariant representation learning techniques including CoRALs, DANNs and DANs.
3.2 free parameters tuning
Free parameters(e.g. learning rate for gradient descent optimizer, layer-wise configurations) selection has been regarded as a crucial role because deep learning methods generally are susceptible to, additionally UDA require us to do this in a agnostic way for target ground truth labels at all. We propose free parameter tuning method specialized in this two stages domain invariant learners by extending RV (same as Normal (Ganin et al., 2017)). RV applies pseudo-labeling to the target data and uses the pseudo-labeled data and the source unsupervised data in an inverse relationship (Zhong et al., 2010), authors proved this can measure conditional distribution difference between target data and learned model. We apply the RV idea to two stages domain invariant representation learners, replacing the internal learning steps from regular supervised machine learning with two stages domain invariant representation learners and calculating classification loss between predictions for source data by reverse model and its ground truth labels. We can find better configurations by . Algorithm is denoted in Algorithm 2, and the Theorem 3.1 states that this method can measure the gap in labeling rules between the trained model and the final target data, the proof of the theorem is given in Appendix B.
Theorem 3.1.
When executing Algorithm 2, conditional distribution gap between and ’s ground truth is calculated by ( as constant values)
(8) |
4 Experimental Validation
4.1 setup and datasets
The evaluation follows the hold-out method. The evaluation score is the percentage of correct labels predicted by the task classifier of the two stages domain invariant learners for the target data i.e. ( as the sample size of target data for testing). The target data is divided into training data and test data in ,the training data is input to the training as unsupervised data, and the test data is not input to the training but used only when calculating the evaluation scores. In order to take into account the variations in the evaluation scores caused by the initial values of each layer of deep learning, 10 evaluations are carried out for each evaluation pattern and the average value is used as the final evaluation score. In addition, hyperparameters optimisation with 3.2 is performed on the learning rate using the training data. The all codes we used in this paper are available on GitHub 111please check v1.0.0(release soon) compatible to this paper [Put URL later], also we elaborated experimental configurations in the Appendix D.
We validate our method with 4 datasets and 38 tasks including simulated toy data, image digits recognition, HAR, occupancy detection. We adopt six benchmark models (conventional Train on Target model, Ste-by-step with DANNs or CoRALs, Normal with DANNs or CoRALs, conventional Without Adapt model) as a comparison test to ours in Table 14 (Appendix I). If our hypothesis is true, our method should be close to its Upper bound and better than previous studies and Lower bound models therefore ideal state ”Upper bound Ours Max(Step-by-step, Normal, Lower bound)” should be expected.
- A. sklearn.datasets.make_moons
-
Source data is two interleaving half circles with binary class. Intermediate data is rotated degrees (semi-clockwise from the centre), target data is rotated degrees (). These rotations correspond to toy versions of two dimensional co-variate shifts. We set noise=0.1. - B. MNIST, MNIST-M, SVHN(Lecun et al., 1998; Ganin et al., 2017; Netzer et al., 2011)
-
Source data is modified national institute of standards and technology database (MNIST), intermediate data is MNIST-M which is the randomly colored version of MNIST, target data is the street view house numbers (SVHN). Case with source and target reversed was demonstrated to be coped with by Normal (Ganin et al., 2017). Two dimensional co-variate shifts should be (Monotone, Not on the Street)(Color, Not on the Street)(Color, On the Street). - C. HHAR(Stisen et al., 2015)
-
Heterogeneity human activity recognition dataset (HHAR). Signal processing problem to determine human behaviour () by accelerometer data (sliding window with size 128, sampling rate is 100-150Hz). Subjects were taking actions in the pre-programmed way, so there are no chances of distribution shift by different or strange movements. The two dimensional co-variate shifts should be (UserA, SensorA)(UserB, SeonsorA)(UserB, SensorB), same actions but different user and sensor differ in co-variate. We extract 16 patterns randomly from 9 users and 4 models. - D. ECO data set(Beckel et al., 2014)
-
Electricity consumption & occupancy (ECO) data set. Signal processing problem same as HHAR with different time window and activity classes(only ). The two dimensional co-variate shifts should be (HouseA, Winter)(HouseB, Winter)(HouseB, Summer). Subjects were not pre-programmed in any means so should include distribution shift in multiple ways, though basically the sharing of the rule of being at home when energy consumption is high and absent when it is not is indicated by (Oshima et al., 2024). 16 Patterns as total.
4.2 quantitative results
We confirmed the Ours’ obvious performance advantage compared to Step-by-step and Normal in Dataset A with DANNs when larger co-variate shift exists i.e. target data with 60 and 70 degrees rotated (highlighted in red in left of Figure 2). The difference was 0.174 compared to Normal when 60 degrees rotated target, 0.074 compared to Step-by-step and 0.091 compared to Normal when 70 degrees rotated target, and variance was much smaller. In Dataset A with CoRALs, increasing the angle of rotation significantly reduces the evaluation scores of Normal, but Step-by-step and Ours were somewhat able to cope with this (highlighted in red in right).
The Figure 3 shows our methods’ superiority to previous studies in Dataset A-D(with DANNs), D(with CoRALs), namely the ideal state ”Upper bound Ours Max(Step-by-step, Normal, Lower bound)” was observed in the UDA experiment. In cases Dataset A and C, a clear difference in accuracy was identified compared to Step-by-step or Normal, the difference in accuracy was 0.074 for Ours and Step-by-step in Dataset C, 0.061 for Ours and Normal, 0.053 for Ours and Normal in Dataset A. In cases other than the above, the degree of deviation from the ideal state is case-by-case. The superiority to Step-by-step i.e. ”Ours Step-by-step” is confirmed 5 out of 8 settings and this demonstrated effectiveness of proposing method i.e. end-to-end domain invariant learning with three sets of data. The superiority to Normal i.e. ”Ours Normal” is confirmed 8 out of 8 settings, highlighting the positive impact of two times domain invariant learnings itself.
Counting the cases where evaluation score is at least higher than the Lower bound, which is important in the UDA setting (”Ours Lower bound”), 8/8 settings. This result means that when UDA is performed in business and real-world settings, a better discriminant model can be built than when it is not performed, highlighting its practical usefulness. Evaluation patterns and results for each Dataset, before aggregation, are provided in the Appendix C.




4.3 qualitative results





















To begin, we investigate how our method performs learning at each epoch in terms of domain invariance and task classifiability. We simultaneously visualise the learning loss of domain invariance per epoch (i.e. ), the learning loss of task classifiability and the synchronised evaluation of task classification on test target data. Figure 4 shows that in any case, domain invariance between the source and the intermediate domain and invariance between the intermediate domain and the target are learnt in an adversarial manner and eventually a solution with high invariance is reached. Also we found that the task classifiability to the source is simultaneously optimised and eventually asymptotically approaches zero or small value. Correspondingly to the above three learnings, the evaluation scores are improving and we can recognise that our proposed optimisation algorithm is effective in the point of task classification performance for target data.
To get more insights into our method, we investigated neural networks’ learned representation at both of feature level and classifier level. The Figure 5 includes feature extractor’s learned representation, task classifier’s sigmoid probability for grid space and representation at feature level is colored in two ways i.e. domain labels and task labels. We can qualitatively recognize ours could learn domain invariant feature between source and target and task discriminability for target data while keeping balance between both. Even though we have not accessed any labels for the target data in black in the figure, we can see a probability boundary that discriminates the data approximately perfectly. On the other hand Normal and Step-by-step could not. Normal could not get domain invariance, task variance, and appropriate boundary. Step-by-step could not learn domain invariance and appropriate boundary.
The Figure 6 shows Ours’ success of domain invariant and task variant features acquirement. Other methods are struggling to get task variant features e.g. the space around the yellow-green scatter points is very dense, including the other three classes. It is difficult for them to classify these human’s behaviour classes (this four classes are far from , Ours could overcome this apparently). The Step-by-step could not learn domain invariant features since we can very easily identify yellow and brown scatter points at a glance.
5 Conclusion and future research direction
We proposed novel UDA strategy whose domain invariance and task variant nature could overcome UDA problem with two dimensional co-variate shifts. Also our proposing free parameters tuning method is useful since it can validate UDA model automatically without access to target ground truth labels.
In terms of research directions for the method, further improvements in accuracy can be expected when the method is used in combination with other UDA methods e.g. (Yang et al., 2024; French et al., 2018; Sun et al., 2022; Yang et al., 2021; Singha et al., 2023). Our method is attractive because it is a broad abstraction that encompasses layers of deep learning and domain invariant representation learning methods internally, and can be used in combination with many other methods (not limited to specific data type e.g. table data, image, and signal). The hyper-parameter optimisation method of (Yang et al., 2024) uses not the conditional distribution gap that can be measured by this method, but with the hopkins statistics (Banerjee & Dave, 2004) and mutual information, transferability at classifier level and transferability and discriminability at feature level can be measured in a combined manner. Also (French et al., 2018) is based on that consistency regularisation ensures that the neural networks’ outputs are close each other for stochastic perturbations. It is well-matched with tons of image augmentation methods of image processing (Shorten & Khoshgoftaar, 2019) and is likely to improve experiments with Dataset B in particular.Safe Self-Refinement for Transformer-based domain adaptation uses vision transformer backbone and consistency regularization as well (Sun et al., 2022), might improve the performance (vision transformer based UDA was also analyzed in another paper (Yang et al., 2021) and improved the performance). Also large vision-language models’ prompting was used for UDA. They format domain invariant and task variant information as the prompting and showed impressive performance improvements in image recognition datasets(Singha et al., 2023).
Another research question is how to obtain an intermediate domain. For time series data, databases normally hold huge amounts of unsupervised data(He et al., 2015; Wang et al., 2019a), so two dimensional co-variate shifts assumption is quite natural. For other data types such as image data, your database can’t always hold a suitable intermediate. One way is to use the Web with huge unlabeled data. Alto this should be a future research direction e.g. generative models possibly will create suitable one based on prompting or papers (Lin et al., 2021; Zhang et al., 2019) methods might help create intermediate synthetically.
References
- Banerjee & Dave (2004) A. Banerjee and R.N. Dave. Validating clusters using the hopkins statistic. In 2004 IEEE International Conference on Fuzzy Systems (IEEE Cat. No.04CH37542), volume 1, pp. 149–153 vol.1, 2004. doi: 10.1109/FUZZY.2004.1375706.
- Beckel et al. (2014) Christian Beckel, Wilhelm Kleiminger, Romano Cicchetti, Thorsten Staake, and Silvia Santini. The eco data set and the performance of non-intrusive load monitoring algorithms. In Proceedings of the 1st ACM International Conference on Embedded Systems for Energy-Efficient Buildings (BuildSys 2014)., pp. 80–89. ACM, 2014.
- Ben-David et al. (2006) Shai Ben-David, John Blitzer, Koby Crammer, and Fernando Pereira. Analysis of representations for domain adaptation. In B. Schölkopf, J. Platt, and T. Hoffman (eds.), Advances in Neural Information Processing Systems, volume 19. MIT Press, 2006.
- Damodaran et al. (2018) Bharath Bhushan Damodaran, Benjamin Kellenberger, Rémi Flamary, Devis Tuia, and Nicolas Courty. Deepjdot: Deep joint distribution optimal transport for unsupervised domain adaptation. In Computer Vision – ECCV 2018: 15th European Conference, Munich, Germany, September 8-14, 2018, Proceedings, Part IV, pp. 467–483, Berlin, Heidelberg, 2018. Springer-Verlag. ISBN 978-3-030-01224-3. doi: 10.1007/978-3-030-01225-0˙28. URL https://doi.org/10.1007/978-3-030-01225-0_28.
- French et al. (2018) Geoff French, Michal Mackiewicz, and Mark Fisher. Self-ensembling for visual domain adaptation. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rkpoTaxA-.
- Ganin et al. (2017) Yaroslav Ganin, Evgeniya Ustinova, Hana Ajakan, Pascal Germain, Hugo Larochelle, François Laviolette, Mario Marchand, and Victor Lempitsky. Domain-Adversarial Training of Neural Networks, pp. 189–209. Springer International Publishing, 2017.
- Gretton et al. (2012) Arthur Gretton, Dino Sejdinovic, Heiko Strathmann, Sivaraman Balakrishnan, Massimiliano Pontil, Kenji Fukumizu, and Bharath K. Sriperumbudur. Optimal kernel choice for large-scale two-sample tests. In F. Pereira, C.J. Burges, L. Bottou, and K.Q. Weinberger (eds.), Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012. URL https://proceedings.neurips.cc/paper_files/paper/2012/file/dbe272bab69f8e13f14b405e038deb64-Paper.pdf.
- He et al. (2015) Guoliang He, Yong Duan, Yifei Li, Tieyun Qian, Jinrong He, and Xiangyang Jia. Active learning for multivariate time series classification with positive unlabeled data. In 2015 IEEE 27th International Conference on Tools with Artificial Intelligence (ICTAI), pp. 178–185, 2015. doi: 10.1109/ICTAI.2015.38.
- Lane et al. (2011) Nicholas D. Lane, Ye Xu, Hong Lu, Shaohan Hu, Tanzeem Choudhury, Andrew T. Campbell, and Feng Zhao. Enabling large-scale human activity inference on smartphones using community similarity networks (csn). In Proceedings of the 13th International Conference on Ubiquitous Computing, UbiComp ’11, pp. 355–364, New York, NY, USA, 2011. Association for Computing Machinery. ISBN 9781450306300. doi: 10.1145/2030112.2030160. URL https://doi.org/10.1145/2030112.2030160.
- Lecun et al. (1998) Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 1998. doi: 10.1109/5.726791.
- Lin et al. (2021) Luojun Lin, Han Xie, Zhishu Sun, Weijie Chen, Wenxi Liu, Yuanlong Yu, and Lei Zhang. Semi-supervised domain generalization with evolving intermediate domain. Pattern Recognit., 149:110280, 2021. URL https://api.semanticscholar.org/CorpusID:257952697.
- Liu et al. (2020) Ziwei Liu, Zhongqi Miao, Xingang Pan, Xiaohang Zhan, Dahua Lin, Stella X. Yu, and Boqing Gong. Open compound domain adaptation. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2020.
- Long et al. (2015) Mingsheng Long, Yue Cao, Jianmin Wang, and Michael Jordan. Learning transferable features with deep adaptation networks. In Francis Bach and David Blei (eds.), Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pp. 97–105, Lille, France, 07–09 Jul 2015. PMLR. URL https://proceedings.mlr.press/v37/long15.html.
- Netzer et al. (2011) Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y. Ng. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and Unsupervised Feature Learning 2011, 2011. URL http://ufldl.stanford.edu/housenumbers/nips2011_housenumbers.pdf.
- Oshima et al. (2024) Hisashi Oshima, Tsuyoshi Ishizone, and Tomoyuki Higuchi. Inter-seasons and inter-households domain adaptation based on danns and pseudo labeling for non-intrusive occupancy detection. The Japanese Society for Artificial Intelligence, 39(5):E–O41_1–13, 2024. URL https://www.jstage.jst.go.jp/article/tjsai/39/5/39_39-5_E-O41/_article.
- Purushotham et al. (2017) Sanjay Purushotham, Wilka Carvalho, Tanachat Nilanon, and Yan Liu. Variational recurrent adversarial deep domain adaptation. In International Conference on Learning Representations, 2017.
- Shorten & Khoshgoftaar (2019) Connor Shorten and Taghi Khoshgoftaar. A survey on image data augmentation for deep learning. Journal of Big Data, 6, 07 2019. doi: 10.1186/s40537-019-0197-0.
- Singha et al. (2023) Mainak Singha, Harsh Pal, Ankit Jha, and Biplab Banerjee. Ad-clip: Adapting domains in prompt space using clip. In 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), pp. 4357–4366, 2023. doi: 10.1109/ICCVW60793.2023.00470.
- Stisen et al. (2015) Allan Stisen, Henrik Blunck, Sourav Bhattacharya, Thor Siiger Prentow, Mikkel Baun Kjærgaard, Anind Dey, Tobias Sonne, and Mads Møller Jensen. Smart devices are different: Assessing and mitigatingmobile sensing heterogeneities for activity recognition. In Proceedings of the 13th ACM Conference on Embedded Networked Sensor Systems, SenSys ’15, pp. 127–140, New York, NY, USA, 2015. Association for Computing Machinery. ISBN 9781450336314. doi: 10.1145/2809695.2809718. URL https://doi.org/10.1145/2809695.2809718.
- Sun & Saenko (2016) Baochen Sun and Kate Saenko. Deep coral: Correlation alignment for deep domain adaptation. In Gang Hua and Hervé Jégou (eds.), Computer Vision – ECCV 2016 Workshops, pp. 443–450, Cham, 2016. Springer International Publishing. ISBN 978-3-319-49409-8.
- Sun et al. (2022) Tao Sun, Cheng Lu, Tianshuo Zhang, and Haibin Ling. Safe self-refinement for transformer-based domain adaptation. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7181–7190, 2022. doi: 10.1109/CVPR52688.2022.00705.
- van der Maaten & Hinton (2008) Laurens van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of Machine Learning Research, 9(86):2579–2605, 2008. URL http://jmlr.org/papers/v9/vandermaaten08a.html.
- Wang et al. (2019a) Haishuai Wang, Qin Zhang, Jia Wu, Shirui Pan, and Yixin Chen. Time series feature learning with labeled and unlabeled data. Pattern Recognition, 89:55–66, 2019a. ISSN 0031-3203. doi: https://doi.org/10.1016/j.patcog.2018.12.026. URL https://www.sciencedirect.com/science/article/pii/S0031320318304473.
- Wang et al. (2017) Yifei Wang, Wen Li, Dengxin Dai, and Luc Van Gool. Deep domain adaptation by geodesic distance minimization. 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), pp. 2651–2657, 2017. URL https://api.semanticscholar.org/CorpusID:4552286.
- Wang et al. (2019b) Zirui Wang, Zihang Dai, Barnabás Póczos, and Jaime Carbonell. Characterizing and avoiding negative transfer. In 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11285–11294, 2019b. doi: 10.1109/CVPR.2019.01155.
- Weiss & Lockhart (2012) Gary Weiss and Jeffrey Lockhart. The impact of personalization on smartphone-based activity recognition. AAAI Workshop - Technical Report, 01 2012.
- Wilson & Cook (2020) Garrett Wilson and Diane J. Cook. A survey of unsupervised deep domain adaptation. ACM Trans. Intell. Syst. Technol., 11(5):1–46, July 2020.
- Wilson et al. (2020) Garrett Wilson, Janardhan Rao Doppa, and Diane J. Cook. Multi-source deep domain adaptation with weak supervision for time-series sensor data. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pp. 1768–1778. ACM, 2020.
- Yang et al. (2024) Jianfei Yang, Hanjie Qian, Yuecong Xu, Kai Wang, and Lihua Xie. Can we evaluate domain adaptation models without target-domain labels? In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview.net/forum?id=fszrlQ2DuP.
- Yang et al. (2021) Jinyu Yang, Jingjing Liu, Ning Xu, and Junzhou Huang. Tvt: Transferable vision transformer for unsupervised domain adaptation. 2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), pp. 520–530, 2021. URL https://api.semanticscholar.org/CorpusID:237048276.
- Zhang et al. (2019) Lei Zhang, Shanshan Wang, Guang-Bin Huang, Wangmeng Zuo, Jian Yang, and David Zhang. Manifold criterion guided transfer learning via intermediate domain generation. IEEE Transactions on Neural Networks and Learning Systems, 30(12):3759–3773, 2019. doi: 10.1109/TNNLS.2019.2899037.
- Zhang et al. (2018a) Yun Zhang, Nianbin Wang, Shaobin Cai, and Lei Song. Unsupervised domain adaptation by mapped correlation alignment. IEEE Access, 6:44698–44706, 2018a. doi: 10.1109/ACCESS.2018.2865249.
- Zhang et al. (2018b) Zhen Zhang, Mianzhi Wang, Yan Huang, and Arye Nehorai. Aligning infinite-dimensional covariance matrices in reproducing kernel hilbert spaces for domain adaptation. In 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3437–3445, 2018b. doi: 10.1109/CVPR.2018.00362.
- Zhao et al. (2019) H. Zhao, Rémi Tachet des Combes, Kun Zhang, and Geoffrey J. Gordon. On learning invariant representations for domain adaptation. In International Conference on Machine Learning(2019), 2019.
- Zhong et al. (2010) Erheng Zhong, Wei Fan, Qiang Yang, Olivier Verscheure, and Jiangtao Ren. Cross validation framework to choose amongst models and datasets for transfer learning. In José Luis Balcázar, Francesco Bonchi, Aristides Gionis, and Michèle Sebag (eds.), Machine Learning and Knowledge Discovery in Databases, pp. 547–562, Berlin, Heidelberg, 2010. Springer Berlin Heidelberg. ISBN 978-3-642-15939-8.
Appendix A CoRALs version two stages domain invariant learners

Appendix B Proof of Reverse Validation based free parameters tuning
Line 2 in Algorithm 2 is saying learned should be approximation of mixing of and line 4 can be seen in the same way. We omitted the descriptions of approximation errors.
Where are the nuisance parameters related to the ratio between the sizes of distributions.
We denoted fixed values as respectively., , ,
Appendix C Quantitative results in detail
Detailed results including the pattern of each data, each method and each domain adaptation task(as PAT for short in these tables) before calculating the average for each dataset and method are in Table 2-13.
PAT | Train on Target | Ours | Step-by-step | Normal(DANNs) | Without Adapt |
---|---|---|---|---|---|
15rotated30rotated | 1(0) | 0.868(.07) | 0.875(.07) | 0.775(.02) | |
20rotated40rotated | 1(0) | 0.869(.08) | 0.873(.06) | 0.619(.03) | |
25rotated50rotated | 1(0) | 0.805(.06) | 0.793(.1) | 0.533(.02) | |
30rotated60rotated | 1(0) | 0.813(.07) | 0.660(.2) | 0.439(.05) | |
35rotated70rotated | 1(0) | 0.700(.1) | 0.683(.2) | 0.339(.01) | |
Average | 1(0) | 0.828(.09) | 0.777(.1) | 0.541(.03) |
PAT | Train on Target | Ours | Step-by-step | Normal(CoRALs) | Without Adapt |
---|---|---|---|---|---|
15rotated30rotated | 1(0) | 0.923(.06) | 0.855(.1) | 0.784(.04) | |
20rotated40rotated | 1(0) | 0.853(.09) | 0.731(.09) | 0.631(.03) | |
25rotated50rotated | 1(0) | 0.633(.1) | 0.560(.07) | 0.529(.03) | |
30rotated60rotated | 1(0) | 0.553(.2) | 0.512(.06) | 0.420(.02) | |
35rotated70rotated | 1(0) | 0.475(.1) | 0.437(.2) | 0.349(.01) | |
Average | 1(0) | 0.687(.1) | 0.619(.08) | 0.543(.03) |
PAT | Train on Target | Ours | Step-by-step | Normal(JDOT) | Without Adapt |
---|---|---|---|---|---|
15rotated30rotated | 1(0) | 0.895(.07) | 0.8882(.04) | 0.782(.04) | |
20rotated40rotated | 1(0) | 0.764(.05) | 0.787(.04) | 0.617(.03) | |
25rotated50rotated | 1(0) | 0.677(.04) | 0.707(.09) | 0.530(.04) | |
30rotated60rotated | 1(0) | 0.522(.03) | 0.554(.1) | 0.427(.02) | |
35rotated70rotated | 1(0) | 0.440(.08) | 0.434(.05) | 0.329(.01) | |
Average | 1(0) | 0.660(.05) | 0.673(.06) | 0.537(.03) |
Trial index | Train on Target | Ours | Step-by-step | Normal(DANNs) | Without Adapt |
---|---|---|---|---|---|
0 | 0.8142286539077759 | 0.2569529712200165 | 0.2764674127101898 | 0.2764674127101898 | |
1 | 0.8248693943023682 | 0.29859402775764465 | 0.29452213644981384 | 0.288836807012558 | |
2 | 0.8270590305328369 | 0.29548248648643494 | 0.2744314670562744 | 0.3106561303138733 | |
Average | 0.8220523596 | 0.302179366350173 | 0.28019360701243 | 0.29198678334554 |
Trial index | Train on Target | Ours | Step-by-step | Normal(CoRALs) | Without Adapt |
---|---|---|---|---|---|
0 | 0.824062705039978 | 0.28626304864883423 | 0.1885755956172943 | 0.2821911573410034 | |
1 | 0.8216425776481628 | 0.267324835062027 | 0.2312154322862625 | 0.23025506734848022 | |
2 | 0.7872233986854553 | 0.231292262673378 | 0.2524969279766083 | 0.20732176303863525 | |
Average | 0.810976227124532 | 0.263137673338254 | 0.20903759698073 | 0.258860886096954 |
Trial index | Train on Target | Ours | Step-by-step | Normal(JDOT) | Without Adapt |
---|---|---|---|---|---|
0 | 0.766364455223083 | 0.252228021621704 | 0.252727419137954 | 0.274469882249832 | |
1 | 0.829706487655639 | 0.281576514244079 | 0.303741544485092 | 0.273547947406768 | |
2 | 0.763560235500335 | 0.23782268166542 | 0.213429629802703 | 0.258681625127792 | |
Average | 0.786543726126352 | 0.264597415924072 | 0.246568332115808 | 0.272126615047455 |
PAT | Train on Target | Ours | Step-by-step | Normal(DANNs) | Without Adapt |
---|---|---|---|---|---|
(d,nexus4)(f,samsungold) | 0.979234308 | 0.333990708 | 0.2445475519 | 0.3040603146 | |
(f,s3mini)(g,s3) | 0.9537375212 | 0.3563122801 | 0.4890365332 | 0.4035714209 | |
(d,s3mini)(e,s3) | 0.9632268667 | 0.6398853391 | 0.6870597839 | 0.8511875629 | |
(b,s3)(f,s3mini) | 0.9091445386 | 0.6380531043 | 0.6666666627 | 0.7166666567 | |
(a,nexus4)(d,s3) | 0.9471365869 | 0.4770925194 | 0.5336563945 | 0.3418502301 | |
(d,s3)(e,samsungold) | 0.9668659866 | 0.3383971155 | 0.3135167331 | 0.2260765433 | |
(e,s3mini)(i,nexus4) | 0.9696132898 | 0.5512707353 | 0.5638305902 | 0.5797790289 | |
(e,samsungold)(f,s3) | 0.9335558951 | 0.4212019861 | 0.4297161847 | 0.3644407243 | |
(f,samsungold)(h,s3) | 0.9565408528 | 0.230904308 | 0.2175592646 | 0.2302897334 | |
(f,s3mini)(g,nexus4) | 0.9656078279 | 0.3412548937 | 0.4018039137 | 0.3747843146 | |
(b,samsungold)(h,s3mini) | 0.8693650961 | 0.2026984192 | 0.2434920691 | 0.2863492191 | |
(c,s3)(i,nexus4) | 0.9692449689 | 0.5139595062 | 0.5067403525 | 0.2813628078 | |
(a,nexus4)(e,s3mini) | 0.8998363554 | 0.512438634 | 0.5076923162 | 0.3450081885 | |
(h,s3)(i,nexus4) | 0.9661510408 | 0.5531123698 | 0.5488398015 | 0.476427266 | |
(b,nexus4)(e,s3mini) | 0.9135843039 | 0.5351882339 | 0.508019653 | 0.4073649794 | |
(a,s3)(b,samsungold) | 0.9768714905 | 0.2344133988 | 0.2617877066 | 0.2089385405 | |
Average | 0.9462323081 | 0.4336480896 | 0.4466236092 | 0.4030455129 |
PAT | Train on Target | Ours | Step-by-step | Normal(CoRALs) | Without Adapt |
(d,nexus4)(f,samsungold) | 0.979234308 | 0.2712296873 | 0.2451276004 | 0.3040603146 | |
(f,s3mini)(g,s3) | 0.9537375212 | 0.5656976521 | 0.5803986609 | 0.4035714209 | |
(d,s3mini)(e,s3) | 0.9632268667 | 0.8188370228 | 0.7868959963 | 0.8158067286 | |
(b,s3)(f,s3mini) | 0.9091445386 | 0.7690265477 | 0.7792035401 | 0.7166666567 | |
(a,nexus4)(d,s3) | 0.9471365869 | 0.5429075032 | 0.529603532 | 0.3418502301 | |
(d,s3)(e,samsungold) | 0.9668659866 | 0.3571770191 | 0.2886363477 | 0.2260765433 | |
(e,s3mini)(i,nexus4) | 0.9696132898 | 0.6311602473 | 0.5859300375 | 0.5797790289 | |
(e,samsungold)(f,s3) | 0.9335558951 | 0.5492487252 | 0.52420699 | 0.3644407243 | |
(f,samsungold)(h,s3) | 0.9565408528 | 0.2807726175 | 0.2603160739 | 0.274363485 | |
(f,s3mini)(g,nexus4) | 0.9656078279 | 0.5189019471 | 0.4879607767 | 0.3747843146 | |
(b,samsungold)(h,s3mini) | 0.8693650961 | 0.2641269937 | 0.2858730286 | 0.2863492191 | |
(c,s3)(i,nexus4) | 0.9692449689 | 0.6111602366 | 0.5978268981 | 0.2813628078 | |
(a,nexus4)(e,s3mini) | 0.8998363554 | 0.5220949322 | 0.5574468166 | 0.3450081885 | |
(h,s3)(i,nexus4) | 0.9661510408 | 0.5809944987 | 0.543167603 | 0.476427266 | |
(b,nexus4)(e,s3mini) | 0.9135843039 | 0.7222586095 | 0.5345335573 | 0.4073649794 | |
(a,s3)(b,samsungold) | 0.9768714905 | 0.2730726153 | 0.2655865803 | 0.2089385405 | |
Average | 0.9462323081 | 0.5197253455 | 0.5120051367 | 0.4030455129 |
PAT | Train on Target | Ours | Step-by-step | Normal(JDOT) | Without Adapt |
---|---|---|---|---|---|
(d,nexus4)(f,samsungold) | 0.979234308 | 0.317865419387817 | 0.321693730354309 | 0.3040603146 | |
(f,s3mini)(g,s3) | 0.9537375212 | 0.438787361979484 | 0.503405302762985 | 0.4035714209 | |
(d,s3mini)(e,s3) | 0.9632268667 | 0.851760858297348 | 0.864209669828415 | 0.8511875629 | |
(b,s3)(f,s3mini) | 0.9091445386 | 0.74144542813301 | 0.73274335861206 | 0.7166666567 | |
(a,nexus4)(d,s3) | 0.9471365869 | 0.389691638946533 | 0.348017627000808 | 0.3418502301 | |
(d,s3)(e,samsungold) | 0.9668659866 | 0.336124384403228 | 0.33349280655384 | 0.2260765433 | |
(e,s3mini)(i,nexus4) | 0.9696132898 | 0.58184163570404 | 0.588397806882858 | 0.5797790289 | |
(e,samsungold)(f,s3) | 0.9335558951 | 0.497829702496528 | 0.527045065164566 | 0.3644407243 | |
(f,samsungold)(h,s3) | 0.9565408528 | 0.146356455236673 | 0.268832318484783 | 0.2808604091 | |
(f,s3mini)(g,nexus4) | 0.9656078279 | 0.332862740755081 | 0.438745093345642 | 0.3747843146 | |
(b,samsungold)(h,s3mini) | 0.8693650961 | 0.287936517596244 | 0.278095249831676 | 0.2863492191 | |
(c,s3)(i,nexus4) | 0.9692449689 | 0.30895028412342 | 0.308692456781864 | 0.2813628078 | |
(a,nexus4)(e,s3mini) | 0.8998363554 | 0.32962357699871 | 0.335024553537368 | 0.3450081885 | |
(h,s3)(i,nexus4) | 0.9661510408 | 0.493407014012336 | 0.518563541769981 | 0.476427266 | |
(b,nexus4)(e,s3mini) | 0.9135843039 | 0.377741411328315 | 0.38936171233654 | 0.4073649794 | |
(a,s3)(b,samsungold) | 0.9768714905 | 0.202234633266925 | 0.258994407951831 | 0.2089385405 | |
Average | 0.9462323081 | 0.427475410280749 | 0.442552631534636 | 0.4030455129 |
PAT | Train on Target | Ours | Step-by-step | Normal(DANNs) | Without Adapt |
---|---|---|---|---|---|
(1, w)(2, s) | 0.8957642913 | 0.7409972489 | 0.7360110939 | 0.6867036104 | |
(1, w)(3, s) | 0.861866653 | 0.7052208841 | 0.6852744341 | 0.6829986632 | |
(2, w)(1, s) | 0.8012861729 | 0.693053323 | 0.6919224679 | 0.6621971011 | |
(2, w)(3, s) | 0.8601333261 | 0.7986613035 | 0.7954484522 | 0.8053547502 | |
(3, w)(1, s) | 0.807395494 | 0.6822294176 | 0.6969305456 | 0.6933764279 | |
(3, w)(2, s) | 0.8927256107 | 0.7096952975 | 0.696583581 | 0.6832871735 | |
(4, w)(5, s) | 0.8287172318 | 0.8516837239 | 0.8380673289 | 0.81859442 | |
(5, w)(4, s) | 0.8698020101 | 0.876616919 | 0.8573797703 | 0.8353233814 | |
(1, s)(2, w) | 0.8868200541 | 0.8035789073 | 0.8214736462 | 0.8545262694 | |
(1, s)(3, w) | 0.780769217 | 0.7954063714 | 0.66077739 | 0.7749116659 | |
(2, s)(1, w) | 0.6769754767 | 0.7716826499 | 0.8065663874 | 0.7937072694 | |
(2, s)(3, w) | 0.7828671217 | 0.7583038926 | 0.7551236808 | 0.760777396 | |
(3, s)(1, w) | 0.724523145 | 0.8347469509 | 0.7835841656 | 0.8228454411 | |
(3, s)(2, w) | 0.8864016414 | 0.8627368033 | 0.8244210184 | 0.8679999769 | |
(4, s)(5, w) | 0.7400809884 | 0.8338086188 | 0.7645621732 | 0.8256619632 | |
(5, s)(4, w) | 0.7781984448 | 0.9102228284 | 0.7644823313 | ||
Average | 0.8171454299 | 0.7820404784 | 0.7924241539 | 0.7754163176 |
PAT | Train on Target | Ours | Step-by-step | Normal(CoRALs) | Without Adapt |
(1, w)(2, s) | 0.8957642913 | 0.7522622466 | 0.7475531042 | 0.6867036104 | |
(1, w)(3, s) | 0.861866653 | 0.7228915513 | 0.7121820569 | 0.6829986632 | |
(2, w)(1, s) | 0.8012861729 | 0.6898223042 | 0.6877221525 | 0.6621971011 | |
(2, w)(3, s) | 0.8601333261 | 0.8285140514 | 0.8327978611 | 0.8078982592 | |
(3, w)(1, s) | 0.807395494 | 0.7008077681 | 0.6957996964 | 0.6933764279 | |
(3, w)(2, s) | 0.8927256107 | 0.7213296473 | 0.7214219868 | 0.6832871735 | |
(4, w)(5, s) | 0.8287172318 | 0.81859442 | |||
(5, w)(4, s) | 0.8698020101 | 0.8825870633 | 0.8812603652 | 0.8353233814 | |
(1, s)(2, w) | 0.8868200541 | 0.831789434 | 0.8061052263 | 0.7663157582 | |
(1, s)(3, w) | 0.780769217 | 0.7176678598 | 0.7975265026 | 0.7749116659 | |
(2, s)(1, w) | 0.6769754767 | 0.7577291667 | 0.7726402462 | 0.7937072694 | |
(2, s)(3, w) | 0.7828671217 | 0.7575971723 | 0.6526501805 | 0.7731448889 | |
(3, s)(1, w) | 0.724523145 | 0.8332421601 | 0.8025992155 | 0.8228454411 | |
(3, s)(2, w) | 0.8864016414 | 0.8532631218 | 0.8418946981 | 0.8774736524 | |
(4, s)(5, w) | 0.7400809884 | 0.8350306153 | 0.8350306153 | 0.8350306153 | |
(5, s)(4, w) | 0.7781984448 | 0.9263434052 | 0.9259502232 | 0.7644823313 | |
Average | 0.8171454299 | 0.7822653914 | 0.7960449468 | 0.7754163176 |
PAT | Train on Target | Ours | Step-by-step | Normal(JDOT) | Without Adapt |
(1, w)(2, s) | 0.8957642913 | 0.73711912035942 | 0.717543876171112 | 0.6867036104 | |
(1, w)(3, s) | 0.861866653 | 0.70575635433197 | 0.696385538578033 | 0.6829986632 | |
(2, w)(1, s) | 0.8012861729 | 0.691599369049072 | 0.686106634140014 | 0.6621971011 | |
(2, w)(3, s) | 0.8601333261 | 0.829183393716812 | 0.810575628280639 | 0.8078982592 | |
(3, w)(1, s) | 0.807395494 | 0.689499223232269 | 0.690468513965606 | 0.6933764279 | |
(3, w)(2, s) | 0.8927256107 | 0.705632507801055 | 0.710249316692352 | 0.6832871735 | |
(4, w)(5, s) | 0.8287172318 | 0.853294265270233 | 0.851098072528839 | 0.81859442 | |
(5, w)(4, s) | 0.8698020101 | 0.866003310680389 | 0.791708122938871 | 0.8353233814 | |
(1, s)(2, w) | 0.8868200541 | 0.82652627825737 | 0.815578907728195 | 0.853052592277526 | |
(1, s)(3, w) | 0.780769217 | 0.74770318865776 | 0.763250893354415 | 0.7749116659 | |
(2, s)(1, w) | 0.6769754767 | 0.786183339357376 | 0.810807144641876 | 0.7937072694 | |
(2, s)(3, w) | 0.7828671217 | 0.768551242351531 | 0.764664322137832 | 0.7763250887 | |
(3, s)(1, w) | 0.724523145 | 0.777428218722343 | 0.808344769477844 | 0.8228454411 | |
(3, s)(2, w) | 0.8864016414 | 0.867789441347122 | 0.855368375778198 | 0.863999962806701 | |
(4, s)(5, w) | 0.7400809884 | 0.838900262117385 | 0.835030615329742 | 0.832179284095764 | |
(5, s)(4, w) | 0.7781984448 | 0.832503297179937 | 0.871428591012954 | 0.7644823313 | |
Average | 0.8171454299 | 0.790123937046155 | 0.787809240445494 | 0.7754163176 |
Appendix D Implementation Configuration in detail
Data load and pre-processing steps are same as UDA previous studies (Ganin et al., 2017; Wilson et al., 2020; Oshima et al., 2024) (experiment with Dataset A and B from (Ganin et al., 2017), Dataset C from (Wilson et al., 2020), Dataset D from (Oshima et al., 2024)). For standardisation pre-processing, the statistics of the target test data are not accessed, only the statistics of the training data are accessed and executed. Internal layers are same between methods, Normal and Step-by-step and Ours have same layers and same shape of respectively though the number of components is not the same since e.g. Normal does not have two domain discriminators. The number of when inference is same. Internal layers are from previous studies, Dataset A with shallow neural networks is from (Ganin et al., 2017), Dataset B with CNN based back bone from (Ganin et al., 2017), Dataset C with one dimensional CNN based from Figure 3 in (Wilson et al., 2020) and Dataset D from Figure 4 and section 4.3. in (Oshima et al., 2024). In the settings during learning, a fixed learning rate is adopted for Dataset A and B. For the other Datasets, the learning rate is determined by optimising from 0.001-0.00001 using the Theorem 3.1 method. In Step-by-step and Normal, the (Ganin et al., 2017) method is used to perform the optimisation. Terminal Evaluation step for target data is exactly same for any method, the feature extractor and task classifier are applied to the 50% of target data not used for training in any sense and their predictions are compared with the ground truth labels to calculate the accuracy. The number of repetitions is three only for Dataset B. The training set of and test set are identical in Dataset A experiment.
Appendix E Two dimensional co-variate shifts Observation
In this section, we confirm whether or not four datasets are following that two dimensional co-variate shifts assumption we introduced in the 2.2. We need to check dataset-wise, (1) existence of marginal distributions shifts between (2) mostly sharing of conditional distribution between any domains. Apparently Dataset A follows that since the three data sets retain co-variate misalignment based on the semi-clockwise rotation action, and without this misalignment the labelling rules are perfectly consistent (Figure 8), likewise in Dataset B can be understood on the action of coloring digit part and background part (Figure 9). About the in Dataset B, although the explicit action does not return to , and includes shifts about labelling rules as well as simple two dimensional co-variate shifts, we speculate a certain sharing of rule based on the facts that we can identify numbers with the human eyes.




Previous work with Dataset C showed their co-variate shift existence between users, so they did the UDA experiment (Wilson et al., 2020) (e.g. see differences between user f and user g in Figure 10). It has been suggested in similar studies (Lane et al., 2011; Weiss & Lockhart, 2012) as well that the characteristics of the data measured will change depending on the age and user of the same behaviour, e.g. ML models trained on data from one age group will not generalise to different age groups. We additionally assume that different models of accelerometer equipment generate further shift based on the analysis by (Stisen et al., 2015) (e.g. see differences between model s3mini and model s3 in Figure 10). The measurement of the same phenomenon under static conditions with different models demonstrated that the distribution of measured values can be very different (please check Figure 1 in (Stisen et al., 2015)). Because all the data for a given model and a given user are taking a predetermined pattern of action classes, we can speculate that is also sharing basically.

It is described by qualitative analysis of the previous study as co-variate shifts that satisfies a certain degree of shared labelling rules in the Dataset D (Oshima et al., 2024). The dynamics include different frequency of electricity peaks for certain time intervals between households (differences in the number of family members) or between seasons, while the previously mentioned rule of being at home if the electricity consumption is large and absent if it is small is shared (please check Figure 2 in (Oshima et al., 2024)). In particular, the Dataset D subjects live in Switzerland, where electricity consumption tends to vary significantly between summer and winter, and is generally higher in winter.
Appendix F Applied researches’ Future Research Direction
A promising direction for applied research is to evaluate the method with data and tasks that are in high demand for other social implementations, and to promote the application of UDA in business and the real world. For example, the problem of determining whether a patient with acute hypoxemic respiratory failure died during hospitalisation from medical data (e.g. blood pH and arterial blood oxygen partial pressure) and the problem of classifying the name of the disease,(Purushotham et al., 2017) may result in distribution shifts due to two dimensional data domains between different ages and different sexes. Semantic image segmentation in self-driving also may put a need for UDA between whether conditions and between a.m. or p.m. e.g. (Noon, Sunny)(Noon, Rainy)(Night, Rainy) (Liu et al., 2020).
Appendix G JDOT version two stages domain invariant learners and experimental validation
Algorithm and schematic diagram are in Algorithm 4 and Figure 11. In the pseudo code, we denote optimal transport solution for sample and sample as for short. Figure 12 shows two-stages JDOT’s superiority to previous studies, namely we found that ”Upper bound Ours Max(Step-by-step, Normal, Lower bound)” for 4 out of 4 datasets. Figure 13 also says that evaluation for ours is always better than Normal and Step-by-step when any rotated target data in Dataset A.



Appendix H Empirical study: effects of two stages domain invariant learners free parameter indicator
We analyzed whether there is a positive correlation between RV-based indicators (left of Theorem 3.1) and actual losses (1st term, right of Theorem 3.1). Using dataset A, learning rate was varied from , and the RV-based score (cross entropy loss) was calculated for the model as the result of UDA learning. At the same time, actual losses (cross entropy loss) were calculated using target data for testing. Finally, pearson correlation coefficient was calculated to determine if there was a positive correlation between the two scores. Table 13 shows that there is a high positive correlation for all patterns. This result supports that two stages domain invariant learners free parameter indicator is useful in UDA experiments.
PAT | 15rotated30rotated | 20rotated40rotated | 25rotated50rotated | 30rotated60rotated | 35rotated70rotated | |||||
---|---|---|---|---|---|---|---|---|---|---|
Method | RV based loss | Ground truth loss | RV based loss | Ground truth loss | RV based loss | Ground truth loss | RV based loss | Ground truth loss | RV based loss | Ground truth loss |
0.00000001 | 0.705 | 0.709 | 0.716 | 0.694 | 0.700 | 0.700 | 0.712 | 0.696 | 0.687 | 0.694 |
0.0000001 | 0.694 | 0.706 | 0.691 | 0.701 | 0.681 | 0.698 | 0.693 | 0.698 | 0.690 | 0.698 |
0.000001 | 0.685 | 0.698 | 0.730 | 0.699 | 0.709 | 0.700 | 0.690 | 0.657 | 0.690 | 0.686 |
0.00001 | 0.692 | 0.683 | 0.706 | 0.612 | 0.687 | 0.664 | 0.713 | 0.950 | 0.693 | 0.689 |
0.0001 | 0.713 | 0.547 | 0.707 | 0.785 | 0.720 | 1.47 | 0.713 | 0.950 | 0.704 | 1.72 |
0.001 | 0.683 | 0.121 | 0.729 | 0.683 | 0.691 | 0.550 | 0.686 | 0.333 | 0.696 | 1.19 |
0.01 | 6.83 | 0.953 | 3.87 | 1.64 | 6.87 | 0.97 | 6.49 | 5.54 | 3.91 | 2.61 |
0.1(learning rate) | 14.7 | 11.7 | 16.1 | 12.6 | 22.3 | 33.5 | 3.65 | 13.4 | 5.05 | 11.8 |
Corr | 0.92 | 0.992 | 0.960 | 0.671 | 0.859 |
Appendix I Benchmarks description for 4 Experimental Validation
We adopted six benchmark models for the comparison test and described input when training, input when inference and what the meaning is when compared to Ours.
Method |
|
|
Role and Note | ||||||
Train on Target | Training data of with its labels | Test data of |
|
||||||
Normal(DANNs,CoRALs) | , training set of | same as above |
|
||||||
Step-by-step |
|
same as above |
|
||||||
Without Adapt | same as above |
|