Abstract
Domain adaptive object detection (DAOD) is an effective approach to solving the domain shift problem, which aims to improve the generalization ability on the target domain through joint training with labeled source domain and unlabeled target domain data. However, in the real world, the source domain is often inaccessible due to privacy, legal, and regulatory reasons. Source-free object detection (SFOD) adapts a well-trained source domain object detector to the target domain without needing access to source domain data. Most existing SFOD methods are based on a self-training framework (such as the teacher–student framework), where the teacher model guides the student model’s training by setting a confidence threshold to filter out unreliable pseudo-labels. However, this may lead to the loss of semantic information in the target domain, resulting in suboptimal model performance. Therefore, in this paper, we propose a contrastive learning framework guided by proposal-level reliable feature guidance. Specifically, we first build a general self-training framework based on Faster-RCNN. Then, we introduce a novel proposal-level pseudo-label annotation and filtering method (PAF), improving the annotation and filtering of pseudo-labels and obtaining proposal-level reliable features. Next, we design reliable feature banks to categorize and store reliable proposal features. Finally, we propose classification contrastive learning (CCL) and bounding box contrastive learning (BCL) to further extract semantic information from reliable proposal features, learning the lost semantic information in the target domain. Extensive experiments on seven benchmark datasets demonstrate the effectiveness of our method.








Similar content being viewed by others
Data availability
No datasets were generated or analysed during the current study
References
Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: Towards real-time object detection with region proposal networks. Adv neural inf process syst 28:1137–1149
Li C, Li L, Jiang H, Weng K, Geng Y, Li L, Ke Z, Li Q, Cheng M, Nie W, et al (2022) Yolov6: a single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976
Redmon J, Divvala S, Girshick R, Farhadi A (2016) You Only Look Once: Unified, Real-time Object Detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788
Torralba A, Efros AA (2011) Unbiased Iook at Dataset Bias. In: CVPR 2011, pp. 1521–1528. IEEE
Lin Y-E, Liu E, Liang X, Chen M, Yan X (2024) Global-local bi-alignment for purer unsupervised domain adaptation. J Supercomput 80(10):14925–14952
Zhang M, Li X, Wu F (2023) Moka-ada: adversarial domain adaptation with model-oriented knowledge adaptation for cross-domain sentiment analysis. J Supercomput 79(12):13724–13743
Hsu C-C, Tsai Y-H, Lin Y-Y, Yang M-H (2020) Every pixel matters: Center-aware Feature Alignment for Domain Adaptive Object Detector. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part IX 16, pp. 733–748. Springer
Chen Y, Li W, Sakaridis C, Dai D, Van Gool L (2018) Domain Adaptive Faster R-cnn for Object Detection in the Wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3339–3348
Deng J, Li W, Chen Y, Duan L (2021) Unbiased Mean Teacher for Cross-domain Object Detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4091–4101
Saito K, Ushiku Y, Harada T, Saenko K (2019) Strong-weak Distribution Alignment for Adaptive Object Detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6956–6965
Inoue N, Furuta R, Yamasaki T, Aizawa K (2018) Cross-domain Weakly-supervised Object Detection Through Progressive Domain Adaptation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5001–5009
Cao S, Joshi D, Gui L-Y, Wang Y-X (2023) Contrastive Mean Teacher for Domain Adaptive Object Detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 23839–23848
Li W, Liu X, Yuan Y (2023) Sigma++: Improved semantic-complete graph matching for domain adaptive object detection. IEEE Trans Pattern Anal Mach Intell 45(7):9022–9040
Vs V, Gupta V, Oza P, Sindagi VA, Patel VM (2021) Mega-cda: Memory Guided Attention for Category-Aware Unsupervised Domain Adaptive Object Detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4516–4526
Chen T, Kornblith S, Norouzi M, Hinton G (2020) A Simple Framework for Contrastive Learning of Visual Representations. In: International Conference on Machine Learning, pp. 1597–1607. PMLR
Johnson-Roberson M, Barto C, Mehta R, Sridhar SN, Rosaen K, Vasudevan R (2016) Driving in the Matrix: Can Virtual Worlds Replace Human-generated Annotations for Real World Tasks? arXiv preprint arXiv:1610.01983
He K, Zhang X, Ren S, Sun J (2016) Deep Residual Learning for Image Recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770–778
Zhang S, Zhang L, Li G, Li P, Liu Z (2023) Multi-prototype guided source-free domain adaptive object detection for autonomous driving. IEEE Trans Intell Veh 9:1589–1601
Sakaridis C, Dai D, Van Gool L (2018) Semantic foggy scene understanding with synthetic data. Int J Comput Vis 126:973–992
Everingham M, Van Gool L, Williams CK, Winn J, Zisserman A (2010) The pascal visual object classes (voc) challenge. Int j comput vis 88:303–338
Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, Franke U, Roth S, Schiele B (2016) The Cityscapes Dataset for Semantic Urban Scene Understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3213–3223
Li S, Ye M, Zhu X, Zhou L, Xiong L (2022) Source-free Object Detection by Learning to Overlook Domain Style. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8014–8023
Xiong L, Ye M, Zhang D, Gan Y, Li X, Zhu Y (2021) Source data-free domain adaptation of object detector through domain-specific perturbation. Int J Intell Syst 36(8):3746–3766
VS V, Oza P, Patel VM (2023) Instance Relation Graph Guided Source-free Domain Adaptive Object Detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3520–3530
Chu Q, Li S, Chen G, Li K, Li X (2023) Adversarial Alignment for Source Free Object Detection. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, pp. 452–460
Huang J, Guan D, Xiao A, Lu S (2021) Model adaptation: historical contrastive learning for unsupervised domain adaptation without source data. Adv neural inf process syst 34:3635–3649
Vs V, Oza P, Sindagi VA, Patel VM (2022) Mixture of Teacher Experts for Source-free Domain Adaptive Object Detection. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3606–3610. IEEE
VS V, Oza P, Patel VM (2023) Towards Online Domain Adaptive Object Detection. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 478–488
Tarvainen A, Valpola H (2017) Mean Teachers are Better Role Models: Weight-averaged Consistency Targets Improve Semi-supervised Deep Learning Results. Adv neural inf process syst 30
Gal Y, Ghahramani Z (2016) Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In: International Conference on Machine Learning, pp. 1050–1059. PMLR
Hsu H-K, Yao C-H, Tsai Y-H, Hung W-C, Tseng H-Y, Singh M, Yang M-H (2020) Progressive Domain Adaptation for Object Detection. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 749–757
Zhu, J-Y, Park T, Isola P, Efros AA (2017) Unpaired Image-to-Image Translation Using Cycle-consistent Adversarial Networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223–2232
Li X, Chen W, Xie D, Yang S, Yuan P, Pu S, Zhuang Y (2021) A Free Lunch for Unsupervised Domain Adaptive Object Detection Without Source Data. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 8474–8481
He K, Fan H, Wu Y, Xie S, Girshick R (2020) Momentum Contrast for Unsupervised Visual Representation Learning. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729–9738
Ouyang F, Shen B (2024) A mutual mean teacher framework for cross-domain aspect-based sentiment analysis. J Supercomput 80(7):9073–9095
Munir MA, Khan MH, Sarfraz M, Ali M (2021) Ssal: Synergizing between self-training and adversarial learning for domain adaptive object detection. Adva Neural Inf Process Syst 34:22770–22782
He Z, Zhang L (2019) Multi-adversarial Faster-rcnn for Unrestricted Object Detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6668–6677
Xu C-D, Zhao X-R, Jin X, Wei X-S (2020) Exploring Categorical Regularization for Domain Adaptive Object Detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11724–11733
He Z, Zhang L (2020) Domain Adaptive Object Detection Via Asymmetric Tri-way Faster-rcnn. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXIV 16, pp. 309–324 Springer
Chen C, Zheng Z, Ding X, Huang Y, Dou Q (2020) Harmonizing Transferability and Discriminability for Adapting Object Detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8869–8878
Chen C, Zheng Z, Huang Y, Ding X, Yu Y (2021) I3net: Implicit Instance-invariant Network for Adapting One-stage Object Detectors. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12576–12585
Geiger A, Lenz P, Stiller C, Urtasun R (2013) Vision meets robotics: The kitti dataset. Int J Robot Res 32(11):1231–1237
Shen Z, Maheshwari H, Yao W, Savvides M (2019) Scl: Towards Accurate Domain Adaptive Object Detection Via Gradient Detach Based Stacked Complementary Losses. arXiv preprint arXiv:1911.02559
Van der Maaten L, Hinton G (2008) Visualizing data using t-sne. J mach learn res 9(11):2579–2605
Acknowledgements
This work was supported by Key Laboratory of AI and Information Processing, Education Department of Guangxi Zhuang Autonomous Region (Hechi University) (2024GXZDSY007), Hefei Natural Science Foundation Project (202308, 202316), Anhui Province Science and Technology Innovation Project (202423k09020003).
Funding
This work was funded by Key Laboratory of AI and Information Processing, Education Department of Guangxi Zhuang Autonomous Region (Hechi University)(2024GXZDSY007), Hefei Natural Science Foundation Project (202308, 202316), Anhui Province Science and Technology Innovation Project (202423k09020003).
Author information
Authors and Affiliations
Contributions
X.J. C. wrote the main manuscript text, Q.J. Z.prepared figures and C.F. completed the experiment.
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wei, X., Xia, J., Liu, C. et al. Proposal-level reliable feature-guided contrastive learning for SFOD. J Supercomput 81, 256 (2025). https://doi.org/10.1007/s11227-024-06773-8
Accepted:
Published:
DOI: https://doi.org/10.1007/s11227-024-06773-8