skip to main content
10.1145/3581783.3611743acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

A Closer Look at Classifier in Adversarial Domain Generalization

Published: 27 October 2023 Publication History

Abstract

The task of domain generalization is to learn a classification model from multiple source domains and generalize it to unknown target domains. The key to domain generalization is learning discriminative domain-invariant features. Invariant representations are achieved using adversarial domain generalization as one of the primary techniques. For example, generative adversarial networks have been widely used, but suffer from the problem of low intra-class diversity, which can lead to poor generalization ability. To address this issue, we propose a new method called auxiliary classifier in adversarial domain generalization (CloCls). CloCls improve the diversity of the source domain by introducing auxiliary classifier. Combining typical task-related losses, e.g., cross-entropy loss for classification and adversarial loss for domain discrimination, our overall goal is to guarantee the learning of condition-invariant features for all source domains while increasing the diversity of source domains. Further, inspired by smoothing optima have improved generalization for supervised learning tasks like classification. We leverage that converging to a smooth minima with respect task loss stabilizes the adversarial training leading to better performance on unseen target domain which can effectively enhances the performance of domain adversarial methods. We have conducted extensive image classification experiments on benchmark datasets in domain generalization, and our model exhibits sufficient generalization ability and outperforms state-of-the-art DG methods.

References

[1]
Yogesh Balaji, Swami Sankaranarayanan, and Rama Chellappa. 2018. Metareg: Towards domain generalization using meta-regularization. NeurIPS, Vol. 31 (2018).
[2]
Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jennifer Wortman Vaughan. 2010. A theory of learning from different domains. Machine learning, Vol. 79, 1 (2010), 151--175.
[3]
Shai Ben-David, John Blitzer, Koby Crammer, and Fernando Pereira. 2006. Analysis of representations for domain adaptation. Advances in neural information processing systems, Vol. 19 (2006).
[4]
Guanyu Cai, Yuqin Wang, Lianghua He, and MengChu Zhou. 2019. Unsupervised domain adaptation with adversarial residual transform networks. IEEE transactions on neural networks and learning systems (TNNLS), Vol. 31, 8 (2019), 3073--3086.
[5]
Fabio M Carlucci, Antonio D'Innocente, Silvia Bucci, Barbara Caputo, and Tatiana Tommasi. 2019a. Domain generalization by solving jigsaw puzzles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2229--2238.
[6]
Fabio Maria Carlucci, Antonio D'Innocente, S. Bucci, B. Caputo, and T. Tommasi. 2019b. Domain Generalization by Solving Jigsaw Puzzles. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019), 2224--2233.
[7]
Junbum Cha, Sanghyuk Chun, Kyungjae Lee, Han-Cheol Cho, Seunghyun Park, Yunsung Lee, and Sungrae Park. 2021. SWAD: Domain Generalization by Seeking Flat Minima. arXiv preprint arXiv:2102.08604 (2021).
[8]
Prithvijit Chattopadhyay, Yogesh Balaji, and Judy Hoffman. 2020. Learning to balance specificity and invariance for in and out of domain generalization. In ECCV. 301--318.
[9]
Chaoqi Chen, Jiongcheng Li, Xiaoguang Han, Xiaoqing Liu, and Yizhou Yu. 2022. Compound Domain Generalization via Meta-Knowledge Encoding. In CVPR. 7119--7129.
[10]
Gintare Karolina Dziugaite and Daniel M Roy. 2017. Computing nonvacuous generalization bounds for deep (stochastic) neural networks with many more parameters than training data. arXiv preprint arXiv:1703.11008 (2017).
[11]
Pierre Foret, Ariel Kleiner, Hossein Mobahi, and Behnam Neyshabur. 2021. Sharpness-aware Minimization for Efficiently Improving Generalization. In International Conference on Learning Representations. https://openreview.net/forum?id=6Tm1mposlrM
[12]
Muhammad Ghifary, W Bastiaan Kleijn, Mengjie Zhang, and David Balduzzi. 2015. Domain generalization for object recognition with multi-task autoencoders. In ICCV. 2551--2559.
[13]
Ishaan Gulrajani and David Lopez-Paz. 2021. In search of lost domain generalization. ICLR (2021).
[14]
Haowei He, Gao Huang, and Yang Yuan. 2019. Asymmetric valleys: beyond sharp and flat local minima. In Proceedings of the 33rd International Conference on Neural Information Processing Systems. 2553--2564.
[15]
Sepp Hochreiter and Jürgen Schmidhuber. 1994. Simplifying neural nets by discovering flat minima. Advances in neural information processing systems, Vol. 7 (1994).
[16]
Sepp Hochreiter and Jürgen Schmidhuber. 1997. Flat minima. Neural computation, Vol. 9, 1 (1997), 1--42.
[17]
Liang Hou, Qi Cao, Huawei Shen, Siyuan Pan, Xiaoshuang Li, and Xueqi Cheng. 2022. Conditional gans with auxiliary discriminative classifier. In International Conference on Machine Learning. PMLR, 8888--8902.
[18]
Zeyi Huang, Haohan Wang, Eric P. Xing, and Dong Huang. 2020a. Self-challenging Improves Cross-Domain Generalization. In European Conference on Computer Vision (ECCV). 124--140.
[19]
Zeyi Huang, Haohan Wang, Eric P Xing, and Dong Huang. 2020b. Self-challenging improves cross-domain generalization. In ECCV. 124--140.
[20]
Nitish Shirish Keskar, Jorge Nocedal, Ping Tak Peter Tang, Dheevatsa Mudigere, and Mikhail Smelyanskiy. 2017. On large-batch training for deep learning: Generalization gap and sharp minima. In 5th International Conference on Learning Representations, ICLR 2017.
[21]
Daehee Kim, Youngjun Yoo, Seunghyun Park, Jinkyu Kim, and Jaekoo Lee. 2021. Selfreg: Self-supervised contrastive regularization for domain generalization. In ICCV. 9619--9628.
[22]
Jungmin Kwon, Jeongseop Kim, Hyunseo Park, and In Kwon Choi. 2021. Asam: Adaptive sharpness-aware minimization for scale-invariant learning of deep neural networks. In International Conference on Machine Learning. PMLR, 5905--5914.
[23]
Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M Hospedales. 2017. Deeper, broader and artier domain generalization. In Proceedings of the IEEE international conference on computer vision. 5542--5550.
[24]
Da Li, Yongxin Yang, Yi-Zhe Song, and Timothy M Hospedales. 2018d. Learning to generalize: Meta-learning for domain generalization. In AAAI.
[25]
Haoliang Li, Sinno Jialin Pan, Shiqi Wang, and Alex C Kot. 2018b. Domain generalization with adversarial feature learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 5400--5409.
[26]
Haoliang Li, Yufei Wang, Renjie Wan, Shiqi Wang, Tie-Qiang Li, and Alex C. Kot. 2020. Domain Generalization for Medical Imaging Classification with Linear-Dependency Regularization. In Advances in Neural Information Processing Systems (NeurIPS).
[27]
Ya Li, Mingming Gong, Xinmei Tian, Tongliang Liu, and Dacheng Tao. 2018a. Domain generalization via conditional invariant representations. In Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 32.
[28]
Ya Li, Xinmei Tian, Mingming Gong, Yajing Liu, Tongliang Liu, Kun Zhang, and Dacheng Tao. 2018c. Deep domain generalization via conditional invariant adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV). 624--639.
[29]
Mingsheng Long, Yue Cao, Jianmin Wang, and Michael Jordan. 2015. Learning transferable features with deep adaptation networks. In International conference on machine learning. PMLR, 97--105.
[30]
Ao Ma, Jingjing Li, Ke Lu, Lei Zhu, and Heng Tao Shen. 2021. Adversarial Entropy Optimization for Unsupervised Domain Adaptation. IEEE Transactions on Neural Networks and Learning Systems (TNNLS) (2021).
[31]
Divyat Mahajan, Shruti Tople, and Amit Sharma. 2020. Domain generalization using causal matching. arXiv preprint arXiv:2006.07500 (2020).
[32]
T. Matsuura and T. Harada. 2020. Domain Generalization Using a Mixture of Multiple Latent Domains. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI).
[33]
Saeid Motiian, Marco Piccirilli, Donald A Adjeroh, and Gianfranco Doretto. 2017. Unified deep supervised domain adaptation and generalization. In ICCV. 5715--5725.
[34]
Krikamol Muandet, David Balduzzi, and Bernhard Schölkopf. 2013. Domain generalization via invariant feature representation. In International Conference on Machine Learning. PMLR, 10--18.
[35]
Hyeonseob Nam, HyunJae Lee, Jongchan Park, Wonjun Yoon, and Donggeun Yoo. 2021. Reducing domain gap by reducing style bias. In CVPR. 8690--8699.
[36]
Sinno Jialin Pan, Ivor W Tsang, James T Kwok, and Qiang Yang. 2010. Domain adaptation via transfer component analysis. IEEE transactions on neural networks, Vol. 22, 2 (2010), 199--210.
[37]
Xingchao Peng, Qinxun Bai, Xide Xia, Zijun Huang, Kate Saenko, and Bo Wang. 2019. Moment matching for multi-source domain adaptation. In Proceedings of the IEEE International Conference on Computer Vision. 1406--1415.
[38]
Vihari Piratla, Praneeth Netrapalli, and Sunita Sarawagi. 2020. Efficient Domain Generalization via Common-Specific Low-Rank Decomposition. In International Conference on Machine Learning (ICML).
[39]
Fengchun Qiao, Long Zhao, and Xi Peng. 2020. Learning to learn single domain generalization. In CVPR. 12556--12565.
[40]
Harsh Rangwani, Sumukh K Aithal, Mayank Mishra, Arihant Jain, and Venkatesh Babu Radhakrishnan. 2022. A closer look at smoothness in domain adversarial training. In International Conference on Machine Learning. PMLR, 18378--18399.
[41]
Seonguk Seo, Yumin Suh, D. Kim, Jongwoo Han, and B. Han. 2020. Learning to Optimize Domain Specific Normalization for Domain Generalization. In European Conference on Computer Vision (ECCV).
[42]
Shiv Shankar, Vihari Piratla, Soumen Chakrabarti, Siddhartha Chaudhuri, Preethi Jyothi, and Sunita Sarawagi. 2018. Generalizing across domains via cross-gradient training. arXiv preprint arXiv:1804.10745 (2018).
[43]
Zhishu Sun, Zhifeng Shen, Luojun Lin, Yuanlong Yu, Zhifeng Yang, Shicai Yang, and Weijie Chen. 2022. Dynamic Domain Generalization. IJCAI (2022).
[44]
Hemanth Venkateswara, Jose Eusebio, Shayok Chakraborty, and Sethuraman Panchanathan. 2017. Deep Hashing Network for Unsupervised Domain Adaptation. In (IEEE) Conference on Computer Vision and Pattern Recognition (CVPR).
[45]
Riccardo Volpi, Hongseok Namkoong, Ozan Sener, John C Duchi, Vittorio Murino, and Silvio Savarese. 2018. Generalizing to unseen domains via adversarial data augmentation. Advances in neural information processing systems, Vol. 31 (2018).
[46]
Mengzhu Wang, Jianlong Yuan, Qi Qian, Zhibin Wang, and Hao Li. 2022. Semantic Data Augmentation based Distance Metric Learning for Domain Generalization. In Proceedings of the 30th ACM International Conference on Multimedia. 3214--3223.
[47]
Shujun Wang, Lequan Yu, Caizi Li, Chi-Wing Fu, and Pheng-Ann Heng. 2020. Learning from extrinsic and intrinsic supervisions for domain generalization. In ECCV. 159--176.
[48]
Zengmao Wang, Bo Du, and Yuhong Guo. 2019. Domain adaptation with neural embedding matching. IEEE transactions on neural networks and learning systems (TNNLS), Vol. 31, 7 (2019), 2387--2397.
[49]
Xiaofu Wu, Suofei Zhang, Quan Zhou, Zhen Yang, Chunming Zhao, and Longin Jan Latecki. 2021. Entropy Minimization Versus Diversity Maximization for Domain Adaptation. IEEE Transactions on Neural Networks and Learning Systems (TNNLS) (2021).
[50]
Qinwei Xu, Ruipeng Zhang, Ya Zhang, Yanfeng Wang, and Qi Tian. 2021. A fourier-based framework for domain generalization. In CVPR. 14383--14392.
[51]
Fu-En Yang, Yuan-Chia Cheng, Zu-Yun Shiau, and Yu-Chiang Frank Wang. 2021. Adversarial teacher-student representation learning for domain generalization. Advances in Neural Information Processing Systems (2021), 19448--19460.
[52]
Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. 2017. mixup: Beyond empirical risk minimization. ICLR (2017).
[53]
Lei Zhang, Jingru Fu, Shanshan Wang, David Zhang, Zhaoyang Dong, and CL Philip Chen. 2019. Guide subspace learning for unsupervised domain adaptation. IEEE transactions on neural networks and learning systems (TNNLS), Vol. 31, 9 (2019), 3374--3388.
[54]
Shanshan Zhao, Mingming Gong, Tongliang Liu, Huan Fu, and Dacheng Tao. 2020a. Domain generalization via entropy regularization. Advances in Neural Information Processing Systems, Vol. 33 (2020), 16096--16107.
[55]
Shanshan Zhao, M. Gong, T. Liu, H. Fu, and Dacheng Tao. 2020b. Domain Generalization via Entropy Regularization. In Advances in Neural Information Processing Systems (NeurIPS).
[56]
Kaiyang Zhou, Yongxin Yang, Timothy Hospedales, and Tao Xiang. 2020a. Deep domain-adversarial image generation for domain generalisation. In Proceedings of the AAAI Conference on Artificial Intelligence. 13025--13032.
[57]
Kaiyang Zhou, Yongxin Yang, Timothy Hospedales, and Tao Xiang. 2020b. Learning to generate novel domains for domain generalization. In ECCV. 561--578.
[58]
Kaiyang Zhou, Yongxin Yang, Yu Qiao, and Tao Xiang. 2021. Domain generalization with mixstyle. ICLR (2021).

Cited By

View all
  • (2025)Gradient-aware domain-invariant learning for domain generalizationMultimedia Systems10.1007/s00530-024-01613-431:1Online publication date: 3-Jan-2025
  • (2024)DGMamba: Domain Generalization via Generalized State Space ModelProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3681247(3607-3616)Online publication date: 28-Oct-2024
  • (2024)Inter-Class and Inter-Domain Semantic Augmentation for Domain GeneralizationIEEE Transactions on Image Processing10.1109/TIP.2024.335442033(1338-1347)Online publication date: 23-Jan-2024
  • Show More Cited By

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
MM '23: Proceedings of the 31st ACM International Conference on Multimedia
October 2023
9913 pages
ISBN:9798400701085
DOI:10.1145/3581783
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 27 October 2023

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. condition-invariant features
  2. domain generalization
  3. smoothing optima

Qualifiers

  • Research-article

Funding Sources

Conference

MM '23
Sponsor:
MM '23: The 31st ACM International Conference on Multimedia
October 29 - November 3, 2023
Ottawa ON, Canada

Acceptance Rates

Overall Acceptance Rate 2,145 of 8,556 submissions, 25%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)179
  • Downloads (Last 6 weeks)4
Reflects downloads up to 19 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Gradient-aware domain-invariant learning for domain generalizationMultimedia Systems10.1007/s00530-024-01613-431:1Online publication date: 3-Jan-2025
  • (2024)DGMamba: Domain Generalization via Generalized State Space ModelProceedings of the 32nd ACM International Conference on Multimedia10.1145/3664647.3681247(3607-3616)Online publication date: 28-Oct-2024
  • (2024)Inter-Class and Inter-Domain Semantic Augmentation for Domain GeneralizationIEEE Transactions on Image Processing10.1109/TIP.2024.335442033(1338-1347)Online publication date: 23-Jan-2024
  • (2024)EDFFF: Ensemble Distillation Framework based on Fourier Features2024 China Automation Congress (CAC)10.1109/CAC63892.2024.10864742(7131-7136)Online publication date: 1-Nov-2024
  • (2024)Architectural Synergies in Bi-Modal and Bi-Contrastive LearningIEEE Access10.1109/ACCESS.2024.345758612(187128-187140)Online publication date: 2024
  • (2023)AI-Driven Health Advice: Evaluating the Potential of Large Language Models as Health AssistantsJournal of Computational Methods in Engineering Applications10.62836/jcmea.v3i1.030106(1-7)Online publication date: 6-Nov-2023
  • (2023)Exploring the Factors of Supply Chain Concentration in Chinese A-Share Listed EnterprisesJournal of Computational Methods in Engineering Applications10.62836/jcmea.v3i1.030105(1-17)Online publication date: 6-Nov-2023

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media