Abstract
In the study of machine learning, multi-source domain adaptation (MSDA) handles multiple datasets which are collected from different distributions by using domain-invariant knowledge extraction. However, the current studies mainly employ features and raw labels on the joint space to perform domain alignment, neglecting the intrinsic structure of label distribution that can harm the performance of adaptation. Therefore, to make better use of label information when aligning joint feature-label distribution, we propose a rebalancing scheme, class-rebalanced Wasserstein distance (CRWD), for unsupervised MSDA under class-wise imbalance and data correlation. Based on the optimal transport for domain adaptation (OTDA) framework, CRWD mitigates the impact of the biased label structure by rectifying the Wasserstein mapping from source to target space. Technically, the class proportions are utilized to encourage distributional transportation between minor classes and principal components, which reweigh the optimal transport plan and reinforce the ground metric of Mahalanobis distance to better metricise the differences among domains. In addition, the scheme measures both inter-domain and intra-source discrepancies to enhance adaptation. Extensive experiments are conducted on various benchmarks, and the results prove that CRWD has competitive advantages.








Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
The data that support the findings of this study are available from the corresponding author, upon reasonable request.
References
Altschuler J, Niles-Weed J, Rigollet P (2017) Near-linear time approximation algorithms for optimal transport via sinkhorn iteration. Adv Neural Inf Process Syst, vol 30
Arora S, Ge R, Liang Y et al (2017) Generalization and equilibrium in generative adversarial nets (gans). In: International conference on machine learning. PMLR, pp 224–232
Brereton RG (2015) The mahalanobis distance and its relationship to principal component scores. J Chemom 29(3):143–145
Cao Y, Long M, Wang J (2018) Unsupervised domain adaptation with distribution matching machines. In: Proceedings of the AAAI conference on artificial intelligence
Chen Q, Liu Y, Wang Z et al (2018) Re-weighted adversarial adaptation network for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7976–7985
Courty N, Flamary R, Tuia D, et al. (2016) Optimal transport for domain adaptation. IEEE Trans Pattern Anal Mach Intell 39(9):1853–1865
Courty N, Flamary R, Habrard A et al (2017) Joint distribution optimal transportation for domain adaptation. Adv Neural Inf Process Syst, vol 30
Cuturi M, Avis D (2014) Ground metric learning. J Mach Learn Res 15(1):533–564
Damodaran BB, Flamary R, Seguy V, et al. (2020) An entropic optimal transport loss for learning deep neural networks under label noise in remote sensing images. Comp Vision Image Underst 191:102,863
David SB, Lu T, Luu T et al (2010) Impossibility theorems for domain adaptation. In: Proceedings of the thirteenth international conference on artificial intelligence and statistics, JMLR workshop and conference proceedings, pp 129–136
Flamary R, Courty N, Gramfort A, et al. (2021) Pot: python optimal transport. J Mach Learn Res 22(78):1–8
Gangbo W, Li W, Osher S et al (2019) Unnormalized optimal transport. J Comput Phys 399:108,940
Gao P, Wu W, Li J (2021) Multi-source fast transfer learning algorithm based on support vector machine. Appl Intell:1–15
Guo J, Shah D, Barzilay R (2018) Multi-source domain adaptation with mixture of experts. In: Proceedings of the 2018 conference on empirical methods in natural language processing, pp 4694–4703
Guo J, Gong M, Liu T et al (2020) Ltf: A label transformation framework for correcting label shift. In: International conference on machine learning, PMLR, pp 3843–3853
Hu C, Wang Y, Gu J (2020) Cross-domain intelligent fault classification of bearings based on tensor-aligned invariant subspace learning and two-dimensional convolutional neural networks. Knowl-Based Syst 209:106–214
Hu C, He S, Wang Y (2021) A classification method to detect faults in a rotating machinery based on kernelled support tensor machine and multilinear principal component analysis. Appl Intell 51 (4):2609–2621
Li Y, Carlson DE, et al. (2018) Extracting relationships by multi-domain matching. Adv Neural Inf Process Syst, vol 31
Liu X, Guo Z, Li S et al (2021) Adversarial unsupervised domain adaptation with conditional and label shift: infer, align and iterate. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 10,367–10,376
Mansour Y, Mohri M, Rostamizadeh A (2008) Domain adaptation with multiple sources. Adv Neural Inf Process Syst:21
Montesuma EF, Mboula FMN (2021) Wasserstein barycenter for multi-source domain adaptation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 16,785–16,793
Peng X, Bai Q, Xia X et al (2019) Moment matching for multi-source domain adaptation. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1406–1415
Peyré G, Cuturi M, et al. (2019) Computational optimal transport: with applications to data science. Foundations Trends Mach Learn 11(5-6):355–607
Podkopaev A, Ramdas A (2021) Distribution-free uncertainty quantification for classification under label shift. In: Uncertainty in artificial intelligence. PMLR, pp 844–853
Rahman MM, Fookes C, Baktashmotlagh M et al (2019) Multi-component image translation for deep domain generalization. In: IEEE winter conference on applications of computer vision (WACV). IEEE, pp 579-588
Rakshit S, Banerjee B, Roig G et al (2019) Unsupervised multi-source domain adaptation driven by deep adversarial ensemble learning. In: German conference on pattern recognition, Springer, pp 485–498
Redko I, Courty N, Flamary R et al (2019) Optimal transport for multi-source domain adaptation under target shift. In: The 22nd international conference on artificial intelligence and statistics. PMLR, pp 849–858
Redko I, Habrard A, Sebban M (2019) On the analysis of adaptability in multi-source domain adaptation. Mach Learn 108(8):1635–1652
Russo P, Tommasi T, Caputo B (2019) Towards multi-source adaptive semantic segmentation. In: International conference on image analysis and processing. Springer, pp 292–301
Saito K, Watanabe K, Ushiku Y et al (2018) Maximum classifier discrepancy for unsupervised domain adaptation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3723–3732
Sun B, Feng J, Saenko K (2017) Correlation alignment for unsupervised domain adaptation. In: Domain adaptation in computer vision applications. Springer, pp 153–171
Sun S, Shi H, Wu Y (2015) A survey of multi-source domain adaptation. Inf Fusion 24:84–92
Turrisi R, Flamary R, Rakotomamonjy A et al (2020) Multi-source domain adaptation via weighted joint distributions optimal transport. arXiv:200612938
Wang H, Xu M, Ni B et al (2020) Learning to combine: knowledge aggregation for multi-source domain adaptation. In: European conference on computer vision. Springer, pp 727– 744
Wang M, Deng W (2018) Deep visual domain adaptation: a survey. Neurocomputing 312:135–153
Wang Z, Jing B, Ni Y, et al. (2020) Adversarial domain adaptation being aware of class relationships. In: ECAI 2020. IOS Press, Santiago de Compostela, pp 1579-1586
Wilson G, Cook DJ (2020) A survey of unsupervised deep domain adaptation. ACM Trans Intell Syst Technol (TIST) 11(5): 1–46
Wu H, Yan Y, Ng MK, et al. (2020) Domain-attention conditional wasserstein distance for multi-source domain adaptation. ACM Trans Intell Syst Technol (TIST) 11(4):1–19
Wu H, Yan Y, Ye Y, et al. (2020) Geometric knowledge embedding for unsupervised domain adaptation. Knowl-Based Syst 191:105,155
Xu R, Chen Z, Zuo W et al (2018) Deep cocktail network: multi-source unsupervised domain adaptation with category shift. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3964–3973
Yang L, Balaji Y, Lim SN et al (2020) Curriculum manager for source selection in multi-source domain adaptation. In: Computer vision–ECCV 2020: 16th European conference, Glasgow, UK, 23–28 August, 2020, proceedings, Part XIV 16. Springer, pp 608–624
Zhang Y, Davison BD (2020) Impact of imagenet model selection on domain adaptation. In: Proceedings of the IEEE/CVF winter conference on applications of computer vision workshops, pp 173–182
Zhao H, Zhang S, Wu G, et al. (2018) Adversarial multiple source domain adaptation. Adv Neural inf process syst 31:8559– 8570
Zhao S, Li B, Yue X, et al. (2019) Multi-source domain adaptation for semantic segmentation. Adv Neural Inf Process Syst, vol 32
Zhao S, Wang G, Zhang S et al (2020) Multi-source distilling domain adaptation. In: Proceedings of the AAAI conference on artificial intelligence, pp 12,975–12,983
Zhao S, Li B, Xu P, et al. (2021) Madan: multi-source adversarial domain aggregation network for domain adaptation. Int J Comput Vis:1–26
Zhu Y, Zhuang F, Wang D (2019) Aligning domain-specific distribution and classifier for cross-domain classification from multiple sources. In: Proceedings of the AAAI conference on artificial intelligence, pp 5989–5996
Zhuang F, Qi Z, Duan K, et al. (2020) A comprehensive survey on transfer learning. Proc IEEE 109(1):43–76
Zuo Y, Yao H, Xu C (2021) Attention-based multi-source domain adaptation. IEEE Trans Image Process 30:3793–3803
Acknowledgements
This work is supported by the Innovation Capacity Construction Project of Jilin Province Development and Reform Commission(2021FGWCXNLJSSZ10), the National Key Research and Development Program of China (No. 2020YFA0714103) and the Science & Technology Development Project of Jilin Province, China (20190302117GX), the Fundamental Research Funds for the Central Universities, JLU..
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Wang, Q., Wang, S. & Wang, B. Class-rebalanced wasserstein distance for multi-source domain adaptation. Appl Intell 53, 8024–8038 (2023). https://doi.org/10.1007/s10489-022-03810-y
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03810-y