Loading [a11y]/accessibility-menu.js
A Parallel Teacher for Synthetic-to-Real Domain Adaptation of Traffic Object Detection | IEEE Journals & Magazine | IEEE Xplore

A Parallel Teacher for Synthetic-to-Real Domain Adaptation of Traffic Object Detection


Abstract:

Large-scale synthetic traffic image datasets have been widely used to make compensate for the insufficient data in real world. However, the mismatch in domain distributio...Show More

Abstract:

Large-scale synthetic traffic image datasets have been widely used to make compensate for the insufficient data in real world. However, the mismatch in domain distribution between synthetic datasets and real datasets hinders the application of the synthetic dataset in the actual vision system of intelligent vehicles. In this paper, we propose a novel synthetic-to-real domain adaptation method to settle the mismatch domain distribution from two aspects, i.e., data level and knowledge level. On the data level, a Style-Content Discriminated Data Recombination (SCD-DR) module is proposed, which decouples the style from content and recombines style and content from different domains to generate a hybrid domain as a transition between synthetic and real domains. On the knowledge level, a novel Iterative Cross-Domain Knowledge Transferring (ICD-KT) module including source knowledge learning, knowledge transferring and knowledge refining is designed, which achieves not only effective domain-invariant feature extraction, but also transfers the knowledge from labeled synthetic images to unlabeled actual images. Comprehensive experiments on public virtual and real dataset pairs demonstrate the effectiveness of our proposed synthetic-to-real domain adaptation approach in object detection of traffic scenes.
Published in: IEEE Transactions on Intelligent Vehicles ( Volume: 7, Issue: 3, September 2022)
Page(s): 441 - 455
Date of Publication: 10 August 2022

ISSN Information:

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.