Skip to main content

Efficient Training Method for Point Cloud-Based Object Detection Models by Combining Environmental Transitions and Active Learning

  • Conference paper
  • First Online:
Robot Intelligence Technology and Applications 7 (RiTA 2022)

Part of the book series: Lecture Notes in Networks and Systems ((LNNS,volume 642))

  • 1142 Accesses

Abstract

The perceptive systems used in automated driving need to function accurately and reliably in a variety of traffic environments. These systems generally perform object detection to identify the positions and attributes of potential obstacles. Among the methods which have been proposed, object detection using three-dimensional (3D) point cloud data obtained using LiDAR has attracted much attention. However, when attempting to create a detection model, annotation must be performed on a huge amount of data. Furthermore, the accuracy of 3D object detection models is dependent on the data domains used for training, such as geographic or traffic environments, so it is necessary to train models for each domain, which requires large amounts of training data for each domain. Therefore, the objective of this study is to develop a 3D object detector for new domains, even when trained with relatively small amounts of annotated data from new domains. We propose using a model that has been trained with a large amount of labeled data for pre-trained model, and simultaneously using transfer learning with limited amount of highly effective training data, selected from the target domain by active learning. Experimental evaluations show that 3D object detection models created using the proposed method perform well at a new location. We also confirm that active learning is particularly effective only limited training data available.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Fagnant, D.J., Kockelman, K.: Preparing a nation for autonomous vehicles: opportunities, barriers and policy recommendations. Transp. Res. Part A: Policy Pract. 77, 167–181 (2015)

    Google Scholar 

  2. Singh, S.: Critical reasons for crashes investigated in the national motor vehicle crash causation survey. Technical report (2015)

    Google Scholar 

  3. Pendleton, S.D., et al.: Perception, planning, control, and coordination for autonomous vehicles. Machines 5(1), 6 (2017)

    Article  Google Scholar 

  4. Arnold, E., Al-Jarrah, O.Y., Dianati, M., Fallah, S., Oxtoby, D., Mouzakitis, A.: A survey on 3D object detection methods for autonomous driving applications. IEEE Trans. Intell. Transp. Syst. 20(10), 3782–3795 (2019)

    Article  Google Scholar 

  5. Chen, Y., Li, W., Sakaridis, C., Dai, D., Van Gool, L.: Domain adaptive faster R-CNN for object detection in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3339–3348 (2018)

    Google Scholar 

  6. Sun, P., et al.: Scalability in perception for autonomous driving: waymo open dataset. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2446–2454 (2020)

    Google Scholar 

  7. Caine, B., et al.: Pseudo-labeling for scalable 3D object detection. arXiv preprint arXiv:2103.02093 (2021)

  8. Pan, S.J., Yang, Q.: A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 22(10), 1345–1359 (2009)

    Article  Google Scholar 

  9. Tan, C., Sun, F., Kong, T., Zhang, W., Yang, C., Liu, C.: A survey on deep transfer learning. In: Kůrková, V., Manolopoulos, Y., Hammer, B., Iliadis, L., Maglogiannis, I. (eds.) ICANN 2018. LNCS, vol. 11141, pp. 270–279. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01424-7_27

    Chapter  Google Scholar 

  10. Zhuang, F.: A comprehensive survey on transfer learning. Proc. IEEE 109(1), 43–76 (2020)

    Article  Google Scholar 

  11. Motional. nuScenes (2021). https://www.nuscenes.org/. Accessed 03 Feb 2022

  12. Caesar, H., et al.: nuScenes: a multimodal dataset for autonomous driving. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11621–11631 (2020)

    Google Scholar 

  13. Citovsky, G., et al.: Batch active learning at scale. Adv. Neural Inf. Process. Syst. 34 (2021)

    Google Scholar 

  14. Ren, P., et al.: A survey of deep active learning. ACM Comput. Surv. (CSUR) 54(9), 1–40 (2021)

    Article  Google Scholar 

  15. Settles, B.: Active learning literature survey. Computer Sciences Technical Report 1648, University of Wisconsin–Madison (2009)

    Google Scholar 

  16. Lang, A.H., Vora, S., Caesar, H., Zhou, L., Yang, J., Beijbom, O.: Pointpillars: fast encoders for object detection from point clouds. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 12697–12705 (2019)

    Google Scholar 

  17. Yin, T., Zhou, X., Krahenbuhl, P.: Center-based 3D object detection and tracking. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11784–11793 (2021)

    Google Scholar 

  18. Geiger, A., Lenz, P., Stiller, C., Urtasun, R.: Vision meets robotics: the KITTI dataset. Int. J. Rob. Res. 32(11), 1231–1237 (2013)

    Article  Google Scholar 

  19. Google. Boston 42\(^\circ \) 20\(^\prime \)45.7398\(^\prime \)N, 71\(^\circ \) 2\(^\prime \)30.0906\(^\prime \)W (2018). https://www.google.co.jp/maps/. Accessed 28 Sept 2021

  20. Google. Singapore 1\(^\circ \) 19\(^\prime \)12.6408\(^\prime \)\(^ \prime \), 103\(^\circ \) 47\(^\prime \)21.8466\(^\prime \)E (2021). https://www.google.co.jp/maps/. Accessed 28 Sept 2021

  21. MMDetection3D (2021). https://github.com/open-mmlab/mmdetection3d. Accessed 09 Dec 2021

  22. Guo, C., Pleiss, G., Sun, Y., Weinberger, K.Q.: On calibration of modern neural networks. In: International Conference on Machine Learning, pp. 1321–1330. PMLR (2017)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Takumi Yamamoto .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yamamoto, T., Ohtani, K., Hayashi, T., Carballo, A., Takeda, K. (2023). Efficient Training Method for Point Cloud-Based Object Detection Models by Combining Environmental Transitions and Active Learning. In: Jo, J., et al. Robot Intelligence Technology and Applications 7. RiTA 2022. Lecture Notes in Networks and Systems, vol 642. Springer, Cham. https://doi.org/10.1007/978-3-031-26889-2_26

Download citation

Publish with us

Policies and ethics