Abstract
In the field of autonomous driving, the importance of the occupancy grid data structure cannot be ignored. The occupancy grid has advantages such as reducing data complexity, improving computational efficiency, and facilitating path planning. By constructing an accurate occupancy grid dataset, researchers can better understand and analyze the distribution of objects in the environment, providing strong support for tasks such as object detection and path planning. This paper proposes a new method for constructing an occupancy dataset, which first constructs dense voxels based on point cloud data, then extracts semantics through two methods, and finally filters the grid based on the visible area to obtain the ground truth of the Occupancy dataset(Named as VA-OCC dataset.). By replacing the existing dataset in the paper with the VA-OCC dataset, better IOU scores and visualization effects can be achieved.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Behley, J., et al.: SemanticKITTI: a dataset for semantic scene understanding of LiDAR sequences. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV) (2019)
Caesar, H., et al.: nuScenes: a multimodal dataset for autonomous driving. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020)
Hackel, T., Savinov, N., Ladicky, L., Wegner, J.D., Schindler, K., Pollefeys, M.: Semantic3D.net: a new large-scale point cloud classification benchmark. In: ISPRS, pp. 91-98 (2017)
Geiger, A., Lenz, P., Urtasun, R.: Are we ready for autonomous driving? The KITTI vision benchmark suite. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition (2012)
Pascal VOC. http://host.robots.ox.ac.uk/pascal/voc/
Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (016)
ADE20K. https://groups.csail.mit.edu/vision/datasets/ade20k/index.html
Wei, Y., Zhao, L., Zheng, W., Zhu, Z., Zhou, J., Lu, J.: SurroundOcc: multi-camera 3D occupancy prediction for autonomous driving. In: ICCV (2023)
Wang, X., et al.: OpenOccupancy : a large scale benchmark for surrounding semantic occupancy perception (2023)
Tian, X., Jiang, T., Yun, L., Wang, Y., Wang, Y., Zhao, H.: Occ3D: a large-scale 3D occupancy prediction benchmark for autonomous driving (2023)
Kirillov, A., et al.: Segment anything. In: ICCV (2023)
Liang, F., et al.: Open-vocabulary semantic segmentation with mask-adapted clip. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7061–7070 (2023)
Xiong, Y., et al.: EfficientSAM: leveraged masked image pretraining for efficient segment anything (2023)
Funding
This work was supported by the Key RD Program in Hubei Province(Grant No. 2022BAA079) and the Key Project of Hubei Province (Grant No. 2021BAA179).
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2025 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Li, Y., Feng, W., Gao, G., Chang, J., Li, M. (2025). VA-OCC : Enhancing Occupancy Dataset Based on Visible Area for Autonomous Driving. In: Antonacopoulos, A., Chaudhuri, S., Chellappa, R., Liu, CL., Bhattacharya, S., Pal, U. (eds) Pattern Recognition. ICPR 2024. Lecture Notes in Computer Science, vol 15317. Springer, Cham. https://doi.org/10.1007/978-3-031-78447-7_24
Download citation
DOI: https://doi.org/10.1007/978-3-031-78447-7_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-78446-0
Online ISBN: 978-3-031-78447-7
eBook Packages: Computer ScienceComputer Science (R0)