Skip to main content

Semantic Scene Filtering for Event Cameras in Long-Term Outdoor Monitoring Scenarios

  • Conference paper
  • First Online:
Advances in Visual Computing (ISVC 2023)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14362))

Included in the following conference series:

  • 335 Accesses

Abstract

Event cameras are biologically inspired devices. They are fundamentally different from conventional frame-based sensors in that they directly transmit an (xyt) output stream of asynchronously and independently detected changes in brightness. For the development of monitoring systems, scenario-based long-term experiments are much more representative than day-to-day experiments. However, unconstrained “real-world” factors pose processing challenges.

To perform a semantic scene filtering on the output stream of an event camera in such an outdoor monitoring scenario, this paper describes a multi-stage processing chain. The goal is to identify and store only those segments that contain events that were triggered by a specific set of objects of interest. The main idea of the proposed processing pipeline is to pre-process the data stream using different filters to identify Patches-Of-Interest (PoIs). These PoIs, natively represented as space-time event clouds, are further processed by PointNet++, a 3D-based semantic segmentation network. An evaluation was performed on about 89 h of real-world outdoor sensor data, achieving a semantic filtering with a false negative rate of \({\approx }3.8\%\) and a true positive rate of \({\approx }96.2\%\).

This work was supported by the European Regional Development Fund under grant number EFRE-0801082 as part of the project “plsm” (https://plsm-project.com/).

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 59.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 79.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    The used CeleX-IV DVS [7] offers a total resolution of \(768\times 640\) px, but due to technical limitations of the sensor hardware, the upper 128 pixel lines were deactivated for recording.

References

  1. Alonso, I., Murillo, A.C.: EV-SegNet: semantic segmentation for event-based cameras. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1624–1633 (2019). https://doi.org/10.1109/CVPRW.2019.00205

  2. Baldwin, R.W., Almatrafi, M., Asari, V., Hirakawa, K.: Event probability mask (EPM) and event denoising convolutional neural network (EDnCNN) for neuromorphic cameras. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020). https://doi.org/10.1109/CVPR42600.2020.00177

  3. Bolten, T., Lentzen, F., Pohle-Fröhlich, R., Tönnies, K.: Evaluation of deep learning based 3D-point-cloud processing techniques for semantic segmentation of neuromorphic vision sensor event-streams. In: Proceedings of the 17th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications - Volume 4: VISAPP, pp. 168–179. INSTICC, SciTePress (2022). https://doi.org/10.5220/0010864700003124

  4. Bolten, T., Pohle-Fröhlich, R., Tönnies, K.D.: DVS-OUTLAB: a neuromorphic event-based long time monitoring dataset for real-world outdoor scenarios. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, pp. 1348–1357 (2021). https://doi.org/10.1109/CVPRW53098.2021.00149

  5. Chen, G., et al.: Multi-cue event information fusion for pedestrian detection with neuromorphic vision sensors. Front. Neurorobot. 13, 10 (2019). https://doi.org/10.3389/fnbot.2019.00010

    Article  Google Scholar 

  6. Gallego, G., et al.: Event-based vision: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(1), 154–180 (2022). https://doi.org/10.1109/TPAMI.2020.3008413

    Article  Google Scholar 

  7. Guo, M., Huang, J., Chen, S.: Live demonstration: a \(768 \times 640\) pixels 200Meps dynamic vision sensor. In: 2017 IEEE International Symposium on Circuits and Systems (ISCAS), p. 1 (2017). https://doi.org/10.1109/ISCAS.2017.8050397

  8. Guo, S., Delbruck, T.: Low cost and latency event camera background activity denoising. IEEE Trans. Pattern Anal. Mach. Intell. (2022). https://doi.org/10.1109/TPAMI.2022.3152999

  9. Guo, S., Wang, L., Chen, X., Zhang, L., Kang, Z., Xu, W.: SeqXFilter: a memory-efficient denoising filter for dynamic vision sensors (2020). https://doi.org/10.48550/arXiv.2006.01687

  10. Jiang, Z., et al.: Mixed frame-/event-driven fast pedestrian detection. In: 2019 International Conference on Robotics and Automation (ICRA), pp. 8332–8338 (2019). https://doi.org/10.1109/ICRA.2019.8793924

  11. Khodamoradi, A., Kastner, R.: O(N)-space spatiotemporal filter for reducing noise in neuromorphic vision sensors. IEEE Trans. Emerg. Top. Comput. 15–23 (2018). https://doi.org/10.1109/TETC.2017.2788865

  12. Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11) (2008)

    Google Scholar 

  13. Perot, E., De Tournemire, P., Nitti, D., Masci, J., Sironi, A.: Learning to detect objects with a 1 megapixel event camera. In: Advances in Neural Information Processing Systems, vol. 33, pp. 16639–16652 (2020). https://doi.org/10.48550/arXiv.2009.13436

  14. Qi, C.R., Yi, L., Su, H., Guibas, L.J.: PointNet++: deep hierarchical feature learning on point sets in a metric space. In: Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 2017, pp. 5105–5114. Curran Associates Inc., Red Hook (2017). https://doi.org/10.48550/arXiv.1706.02413

  15. Sabater, A., Montesano, L., Murillo, A.C.: Event transformer. a sparse-aware solution for efficient event data processing. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 2677–2686 (2022). https://doi.org/10.1109/CVPRW56347.2022.00301

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tobias Bolten .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Bolten, T., Pohle-Fröhlich, R., Tönnies, K.D. (2023). Semantic Scene Filtering for Event Cameras in Long-Term Outdoor Monitoring Scenarios. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2023. Lecture Notes in Computer Science, vol 14362. Springer, Cham. https://doi.org/10.1007/978-3-031-47966-3_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-031-47966-3_7

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-031-47965-6

  • Online ISBN: 978-3-031-47966-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics