Skip to main content
Log in

Litter Detection from Digital Images Using Deep Learning

  • Original Research
  • Published:
SN Computer Science Aims and scope Submit manuscript

Abstract

In order to achieve automatically litter detection in residential area, machine vision has been applied to monitor environment of surveillance. Based on our observations and comparative analysis of the current algorithms, we propose an improved object detection method based on Faster R-CNN algorithm and achieve more than 98% accuracy of litter detection in surveillance. Through our observations, most of litters are small objects, we apply feature pyramid network to Faster R-CNN and optimize it by merging different layers by using multiply operate. Besides, we replace cross-entropy loss function with focal loss function to solve the problem of anchor imbalance by using region proposal network (RPN) and offer attention module through RPN to feedback the whole network. We collected more than 8000 labeled images from our surveillance videos for model training. Our experiments show that the improved Faster R-CNN achieves a satisfied performance in real scene.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Data availability

No data was used for the research described in the article.

References

  1. Bochkovskiy A, Wang CY, Liao HM. YOLOv4: optimal speed and accuracy of object detection. IEEE CVPR; 2020.

  2. Cai Z, Fan Q, Feris RS, Vasconcelos N. A unified multi-scale deep convolutional neural network for fast object detection. European Conference on Computer Vision; 2016. p. 354–370.

  3. De Boer PT, Kroese DP, Mannor S, Rubinstein RY. A tutorial on the cross-entropy method. Ann Oper Res. 2005;134(1):19–67.

    Article  MathSciNet  MATH  Google Scholar 

  4. Girshick R. Fast R-CNN. International conference on computer vision (ICCV). Santiago, 2015; 2015. p. 1440–1448.

  5. He K, Gkioxari G, Dollár P, Girshick R. Mask R-CNN. IEEE international conference on computer vision; 2017. p. 2961–2969.

  6. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. IEEE conference on computer vision and pattern recognition; 2016. p. 770–778.

  7. Heikkila M, Pietikainen M. A texture-based method for modeling the background and detecting moving objects. IEEE Trans Pattern Anal Mach Intell. 2006;28(4):657–62.

    Article  Google Scholar 

  8. Kong T, Yao A, Chen Y, Sun F. HyperNet: towards accurate region proposal generation and joint object detection. IEEE conference on computer vision and pattern recognition; 2016.

  9. Krizhevsky A, Sutskever I, Hinton GE. ImageNet classification with deep convolutional neural networks. Neural Information Processing Systems; 2012.

  10. Lee SH, Yeh CH, Hou TW, Yang CS. A lightweight neural network based on AlexNet-SSD model for garbage detection. High performance computing and cluster technologies conference; 2019. p. 274–278.

  11. Li L, Su H, Feifei L, Xing EP. Object bank: a high-level image representation for scene classification and semantic feature sparsification. Neural Information Processing Systems; 2010.

  12. Lin TY, Dollár P, Girshick R, He K, Hariharan B, Belongie S. Feature pyramid networks for object detection. IEEE CVPR; 2017. p. 2117–2125.

  13. Lin T, Goyal P, Girshick R, He K, Dollar P. Focal loss for dense object detection. IEEE ICCV; 2017.

  14. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, Berg AC. SSD: single shot multibox detector. European conference on computer vision; 2016. p. 21–37.

  15. Long J, Shelhamer E, Darrell T. Fully convolutional networks for semantic segmentation. International conference on computer vision and pattern recognition; 2015.

  16. Ma X, Chen Z, Zhang J. Fully convolutional network with cluster for semantic segmentation. International conference on AMME; 2018.

  17. Marr D, Hildreth EC. Theory of edge detection. Proc R Soc B Biol Sci. 1980;207(1167):187–217.

    Google Scholar 

  18. Mittal G, Yagnik KB, Garg M, Krishnan NC. SpotGarbage: smartphone App to detect garbage using deep learning. ACM international joint conference on pervasive and ubiquitous computing; 2016. p. 940–945.

  19. Qin Z, Li Z, Zhang Z, Bao Y, Yu G, Peng Y, Sun J. ThunderNet: towards real-time generic object detection. International conference on computer vision and pattern recognition. 2019.

  20. Redmon J, Farhadi A. YOLOv3: an incremental improvement. 2018. arXiv:1804.02767.

  21. Redmon J, Farhadi A. YOLO9000: better, faster, stronger. IEEE conference on computer vision and pattern recognition; 2017. p. 7263–7271.

  22. Ren S, He K, Girshick R, Sun J. Faster R-CNN: towards real-time object detection with region proposal networks. Advances in neural information processing systems; 2015. p. 91–99.

  23. Romeraparedes B, Torr PH. Recurrent instance segmentation. European conference on computer vision; 2016. p. 312–329.

  24. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014. arXiv:1409.1556.

  25. Thakare BS, Dube MR. New approach for model merging and transformation. International conference on computer communication and informatics; 2012.

  26. Tian Z, Shen C, Chen H, He T. FCOS: fully convolutional one-stage object detection. IEEE conference on computer vision and pattern recognition; 2019.

  27. Zhou Z, Shin J, Zhang L, Gurudu S, Gotway M, Liang J. Fine-tuning convolutional neural networks for biomedical image analysis: actively and incrementally. IEEE conference on computer vision and pattern recognition; 2017. p. 7340–7351.

  28. Uijlings JR, Van De Sande KE, Gevers T, Smeulders AW. Selective search for object recognition. Int J Comput Vis. 2013;104(2):154–71.

    Article  Google Scholar 

  29. Wang Y, Zhang X. Autonomous garbage detection for intelligent urban management. MATEC Web of Conferences, vol. 232. EDP Sciences; 2018. p. 01056.

  30. Xie Y, Chen Y. Object tracking based on spatial attention mechanism. Chinese control conference; 2019.

  31. Zhong Y, Wang J, Peng J, Zhang L. Anchor box optimization for object detection. IEEE winter conference on applications of computer vision; 2020. p. 1275–1283.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wei Qi Yan.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

This article is part of the topical collection “From Geometry to Vision: The Methods for Solving Visual Problems” guest edited by Wei Qi Yan, Harvey Ho, Minh Nguyen and Zhixun Su.

Appendix

Appendix

The notation or abbreviation table for the symbols

CE loss

Cross-entropy loss

FPN

Feature pyramid network

R-CNN

Region-based convolutional neural network

RoI

Region of interest

RPN

Region proposal network

SSD

Single shot multibox detector

VGG

Very deep convolutional networks

YOLO

You only look once

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Liu, J., Pan, C. & Yan, W.Q. Litter Detection from Digital Images Using Deep Learning. SN COMPUT. SCI. 4, 134 (2023). https://doi.org/10.1007/s42979-022-01568-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s42979-022-01568-1

Keywords

Navigation