Abstract
Recently, object recognition and image segmentation have gained much attention in the computer vision field and image processing for effective object localisation and identification. Researchers have applied semantic segmentation and instance segmentation in diverse application areas. However, the least research has been performed in natural habitat monitoring or plant species identification in natural environments/surroundings. For this study, we composed a real image dataset from four habitats: forests, dunes, grasslands, and screes from various locations in Italy. Habitat expert botanists annotated the data using bounding box annotations which have been further utilised to generate the plant species masks using the recently proposed Segment Anything Model (SAM) for segmentation, localisation, and identification tasks. Extensive experimentation has been performed on habitat data with bounding boxes and masks using YOLOv8 detection and segmentation models. Comparative analysis of models, model training with different train data percentages, and the importance of masks over bounding boxes have been studied and discussed.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Angelini, F., et al.: Robotic monitoring of habitats: the natural intelligence approach. IEEE Access (2023)
Arya, S., Sandhu, K.S., Singh, J., Kumar, S.: Deep learning: as the new frontier in high-throughput plant phenotyping. Euphytica 218(4), 47 (2022)
Deng, R., et al.: Segment anything model (SAM) for digital pathology: assess zero-shot segmentation on whole slide imaging. arXiv preprint arXiv:2304.04155 (2023)
Dosovitskiy, A., et al.: An image is worth \(16 \times 16\) words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)
Goëau, H., Bonnet, P., Joly, A.: Overview of plantCLEF 2022: image-based plant identification at global scale. In: CLEF 2022-Conference and Labs of the Evaluation Forum, vol. 3180, pp. 1916–1928 (2022)
He, K., Chen, X., Xie, S., Li, Y., Dollár, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000–16009 (2022)
Huang, Y., et al.: Segment anything model for medical images? arXiv preprint arXiv:2304.14660 (2023)
Jocher, G.: Ultralytics YOLOv8 github (2023). https://github.com/ultralytics/ultralytics
Jocher, G., et al.: ultralytics/YOLOv5: v7. 0-YOLOv5 SOTA realtime instance segmentation. Zenodo (2022)
Kang, J., Zhao, L., Wang, K., Zhang, K., et al.: Research on an improved YOLOv8 image segmentation model for crop pests. Adv. Comput. Signals Syst. 7(3), 1–8 (2023)
Kirillov, A., et al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)
Labelbox: Labelbox (2023). https://labelbox.com
Mikula, K., et al.: Naturasat—a software tool for identification, monitoring and evaluation of habitats by remote sensing techniques. Remote Sens. 13(17), 3381 (2021)
Pushpa, B., Rani, N.: Ayur-PlantNet: an unbiased light weight deep convolutional neural network for Indian ayurvedic plant species classification. J. Appl. Res. Med. Aromatic Plants 34, 100459 (2023)
Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016)
Redmon, J., Farhadi, A.: YOLOv3: an incremental improvement. arXiv preprint arXiv:1804.02767 (2018)
Sun, Y., Liu, Y., Wang, G., Zhang, H., et al.: Deep learning for plant identification in natural environment. Comput. Intell. Neurosci. 2017 (2017)
Ultralytics Team: Ultralytics YOLOv8 docs (2023). https://docs.ultralytics.com/
Ultralytics: Ultralytics YOLOv8 (2023). https://github.com/ultralytics/ultralytics/issues/189
Williams, D., MacFarlane, F., Britten, A.: Leaf only SAM: a segment anything pipeline for zero-shot automated leaf segmentation. arXiv preprint arXiv:2305.09418 (2023)
Xu, F., Li, B., Xu, S.: Accurate and rapid localization of tea bud leaf picking point based on YOLOv8. In: Meng, X., Chen, Y., Suo, L., Xuan, Q., Zhang, Z.K. (eds.) BDSC 2023. CCIS, vol. 1846, pp. 261–274. Springer, Singapore (2023). https://doi.org/10.1007/978-981-99-3925-1_17
Xu, M., Yoon, S., Jeong, Y., Lee, J., Park, D.S.: Transfer learning with self-supervised vision transformer for large-scale plant identification. In: International Conference of the Cross-Language Evaluation Forum for European Languages, pp. 2253–2261. Springer (2022)
Yan, B., Fan, P., Lei, X., Liu, Z., Yang, F.: A real-time apple targets detection method for picking robot based on improved YOLOv5. Remote Sens. 13(9), 1619 (2021)
Yu, T., et al.: Inpaint anything: segment anything meets image inpainting. arXiv preprint arXiv:2304.06790 (2023)
Yuan, Q., et al.: Deep learning in environmental remote sensing: achievements and challenges. Remote Sens. Environ. 241, 111716 (2020)
Zhao, H., et al.: Jujube fruit instance segmentation based on yolov8 method. Available at SSRN 4482151 (2023)
Acknowledgement
This research was supported by Grant agreement No. 101016970, European Union’s Horizon 2020 Research and Innovation Programme - ICT-47-2020.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kaur, P. et al. (2023). Segmentation and Identification of Mediterranean Plant Species. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2023. Lecture Notes in Computer Science, vol 14362. Springer, Cham. https://doi.org/10.1007/978-3-031-47966-3_34
Download citation
DOI: https://doi.org/10.1007/978-3-031-47966-3_34
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-47965-6
Online ISBN: 978-3-031-47966-3
eBook Packages: Computer ScienceComputer Science (R0)