Skip to main content

Advertisement

Log in

Rapid density estimation of tiny pests from sticky traps using Qpest RCNN in conjunction with UWB-UAV-based IoT framework

  • S.I.: Neural Networks and Machine Learning Empowered Methods and Applications in Healthcare
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Precision agriculture has long struggled with the surveillance and control of pests. Traditional methods for estimating pest density and distribution through manual reconnaissance are often time-consuming and labor-intensive. To address these challenges, this study proposes a novel farmland detection system utilizing UAVs, yellow sticky traps, and deep learning techniques. The system includes a UAV based on UWB communication and positioning technology to collect picture information of sticky traps arranged in farmland. Moreover, a faster and more accurate Qpest-RCNN model is used to count the number of insects in the collected sticky traps. The bivariate kernel density estimation establishes the pest density distribution map. Regarding the dynamic monitoring of pest density in agricultural fields, it primarily involves four components: reaching the designated area, flight learning, image acquisition, and visual counting of insects and calculation of insect density. Experimental results demonstrate that UAVs require less time to adjust flight posture during image acquisition after undergoing flight learning, resulting in more concise flight trajectories. The Qpest-RCNN model introduces a variety of mechanisms based on the characteristics of the collected sticky trap data set to improve the faster-R-CNN model. We used two data sets separately to train the model, which were collected by sticky traps placed in greenhouses and open-air experimental fields. The data set from the greenhouse is an open-source dataset provided by M. Deserno et. al, map, precision, and recall of model are 0.923, 0.989, and 0.919, respectively. When the data set collected in the experimental farmland is used to train the model, map, precision, and recall of model are 0.781, 0.851, and 0.789, respectively. In the meantime, we explored the effects of species interference on visual insect statistics and the optimization effect of species hypothesis on statistics in two environments. At the same time, the inference speed of the improved model is about a quarter faster than the FPS of the original Faster-RCNN during inference. Through Qpest-RCNN, the number of insects captured on sticky traps can be counted quickly and accurately and the bivariate kernel density estimation is used to draw the pest density distribution map to observe the pest distribution of the whole farmland visually. This pest-density farmland detection system is valuable for agriculture by helping farmers control pests, reduce crop damage, and increase yield and quality.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data availability

The data that support the findings of this study are available from the corresponding author upon reasonable request.

References

  1. Shin YK, Kim SB, Kim D-SJJOA-PE (2020) Attraction characteristics of insect pests and natural enemies according to the vertical position of yellow sticky traps in a strawberry farm with high-raised bed cultivation. J Asia-Pacific Entomol 23(4):1062–1066

    Article  MathSciNet  Google Scholar 

  2. Pinto-Zevallos DM, Vänninen IJCP (2013) Yellow sticky traps for decision-making in whitefly management: what has been achieved? Crop Prot 47:74–84

    Article  Google Scholar 

  3. Qiao M et al (2008) Density estimation of Bemisia tabaci (Hemiptera: Aleyrodidae) in a greenhouse using sticky traps in conjunction with an image processing system. J Asia-Pac Entomol 11(1):25–29

    Article  Google Scholar 

  4. Hall DG, Sétamou M, Mizell RFJCP III (2010) A comparison of sticky traps for monitoring Asian citrus psyllid (Diaphorina citri Kuwayama). Elsevier 29(11):1341–1346

    Google Scholar 

  5. Rodríguez LAR, Castañeda-Miranda CL, Lució MM, Solís-Sánchez LO, Castañeda-Miranda RJM, Simulation CI (2021) Quarternion color image processing as an alternative to classical grayscale conversion approaches for pest detection using yellow sticky traps. Math Comput Simul 182:646–660

    Article  MathSciNet  Google Scholar 

  6. Sun Y et al (2017) A smart-vision algorithm for counting whiteflies and thrips on sticky traps using two-dimensional Fourier transform spectrum. Biosys Eng 153:82–88

    Article  Google Scholar 

  7. Lee H, Choi W, Eom S, Park J-JJJOA-PB (2022) "Rapid estimation of the density of whiteflies (Hemiptera: Aleyrodidae) on sticky traps in paprika greenhouses using the presence–absence model. J Asia-Pac Biodiv 15(2):225–230

    Google Scholar 

  8. Li W et al (2021) Field detection of tiny pests from sticky trap images using deep learning in agricultural greenhouse. Comput Electron Agric 183:106048

    Article  Google Scholar 

  9. Boursianis AD et al (2022) Internet of things (IoT) and agricultural unmanned aerial vehicles (UAVs) in smart farming: a comprehensive review. Internet of Things 18:100187

    Article  Google Scholar 

  10. Srivastava K, Bhutoria AJ, Sharma JK, Sinha A, PCJRSAS Pandey and Environment (2019) UAVs technology for the development of GUI based application for precision agriculture and environmental research. Remote Sens Appl: Soc Environ 16:100258

    Google Scholar 

  11. Anand S, Sharma AJMS (2022) AgroKy: an approach for enhancing security services in precision agriculture. Meas: Sens 24:100449

    Google Scholar 

  12. CouliablyS Kamsu-Foguem B, Kamissoko D, Traore DJISwA, (2022) Deep learning for precision agriculture: A bibliometric analysis. Intel Syst Appl 16:200102

    Google Scholar 

  13. Cui M, Qian J, LJJoRS Cui (2022) Developing precision agriculture through creating information processing capability in rural China. J Rural Stud 92:237–252

    Article  Google Scholar 

  14. Hanson ED, Cossette MK, Roberts DCJTiS (2022) The adoption and usage of precision agriculture technologies in North Dakota. Tech Soc 71:102087

    Article  Google Scholar 

  15. Ampatzidis Y, Partel V, Costa LJ (2020) Agroview: Cloud-based application to process, analyze and visualize UAV-collected data for precision agriculture applications utilizing artificial intelligence. Comput Electron Agri 174:105457

    Article  Google Scholar 

  16. Comba L, Biglia A, Aimonino DR, Gay PJC, EI Agriculture (2018) Unsupervised detection of vineyards by 3D point-cloud UAV photogrammetry for precision agriculture. Comput Electron Agric 155(84):95

    Google Scholar 

  17. Elmokadem TJI-P (2019) Distributed coverage control of quadrotor multi-UAV systems for precision agriculture. IFAC-PapersOnLine 52(30):251–256

    Article  Google Scholar 

  18. Radoglou-Grammatikis P, Sarigiannidis P, Lagkas T, Moscholios IJCN (2020) A compilation of UAV applications for precision agriculture. Comput Netw 172:107148

    Article  Google Scholar 

  19. Su J, Zhu X, Li S, Chen W-HJN (2022) AI meets UAVs: A survey on AI empowered UAV perception systems for precision agriculture. Neurocomputing 518:242–270

    Article  Google Scholar 

  20. Singh PK, Sharma AJC, Engineering E (2022) An intelligent WSN-UAV-based IoT framework for precision agriculture application. Comput Electr Eng 100:107912

    Article  Google Scholar 

  21. Tetila EC et al (2020) Detection and classification of soybean pests using deep learning with UAV images. Comput Electron Agric 179:105836

    Article  Google Scholar 

  22. Nieuwenhuizen AT, Hemming J, Suh HK (2018) Detection and classification of insects on stick-traps in a tomato crop using Faster R-CNN. In: The Netherlands Conference on Computer Vision

  23. Nieuwenhuizen A et al (2019) Raw data from Yellow Sticky Traps with insects for training of deep learning convolutional neural network for object detection

  24. DesernoM, Briassouli A (2021) Faster r-CNN and efficientnet for accurate insect identification in a relabeled yellow sticky traps dataset. In: 2021 IEEE international workshop on metrology for agriculture and forestry (MetroAgriFor), 2021: IEEE, pp 209–214

  25. Mo H, Zhang J, Ma Y (2019) Application and system design of dw1000 in UAV cluster. Electron World 2019(12):135–137

    Google Scholar 

  26. Zheng Y, Xue L, Dong L (2019) Formation control of mobile robots with UWB localization technology. Chin J Intell Sci Technol 1(1):83–87

    Google Scholar 

  27. Liu B, Zhang R, Li Y, Wang Y, Fu P (2019) An UWB-based wireless module set with transparent transmission. J China Acad Electron Inf Technol 14(2):168–176

    Google Scholar 

  28. Shi D, Liu C, She F (2022) Cooperation localization method based on location confidence of multi-UAV in GPS-denied environment. Comput Sci 49(4):302–311

    Google Scholar 

  29. Yu S (2020) Design and realization of flight control for UAV Group’s autonomous formation. Inner Mongolia University, Hohhot

    Google Scholar 

  30. HuZ Tong Q, Liu SJJCJU (2020) Lane departure identification and early warning method for autonomous vehicle. J Chongqing Jiaotong Univer (Nat Sci) 39(10):118–125

    Google Scholar 

  31. Wang L (2017) Research and implementation of car lane departure warning system based on image processing. University of Electronic Science and Technology of China

  32. Li G (2021) Design of UAV navigation system based on image recognition. J Agric Mech Res 43(1):114–118

    Google Scholar 

  33. Ibrahim MH (2021) ODBOT: outlier detection-based oversampling technique for imbalanced datasets learning. Neural Comput Appl 33(22):15781–15806

    Article  Google Scholar 

  34. RenS, He K, Girshick R, Sun JJA i n i p s (2015) Faster r-cnn: toward real-time object detection with region proposal networks. vol 28

  35. Girshick R (2015) Fast r-cnn. In: Proceedings of the IEEE international conference on computer vision, pp 1440–1448

  36. Lin T-Y, Dollár P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117–2125

  37. Yang C, Huang Z, Wang N (2022) QueryDet: cascaded sparse query for accelerating high-resolution small object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 13668–13677

  38. Lin T-Y, Goyal P, Girshick R, He K, Dollár P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980–2988

  39. BodlaN, Singh B, Chellappa R, Davis LS (2017) Soft-NMS-improving object detection with one line of code. In: Proceedings of the IEEE international conference on computer vision, 2017, pp 5561–5569

  40. Sønderby SK, Sønderby CK, Maaløe L, Winther OJapa (2015) Recurrent spatial transformer networks. arXiv preprint arXiv:1509.05329

  41. Xu Z (2021) Research on traffic target detection algorithm based on improved faster RCNN. https://doi.org/10.26976/dcnki.Gchau.2021.000610

  42. ZhuX, Cheng D, Zhang Z, Lin S, Dai J (2019) An empirical study of spatial attention mechanisms in deep networks. In: Proceedings of the IEEE/CVF international conference on computer vision, pp. 6688–6697

  43. Simonoff JS (2012) Smoothing methods in statistics. Springer, Berlin

    Google Scholar 

  44. Botev ZI, Grotowski JF, Kroese DP (2010) Kernel density estimation via diffusion. Ann Stat 38(5):2916–2957

    Article  MathSciNet  Google Scholar 

  45. Cheng C (2022) Real-time mask detection based on SSD-MobileNetV2. In: 2022 IEEE 5th international conference on automation, electronics and electrical engineering (AUTEEE), IEEE, pp 761–767

  46. Redmon J, Farhadi AJapa (2018) Yolov3: an incremental improvement. arXiv e-prints

Download references

Acknowledgements

Supported by the National Natural Science Foundation of China (61771034). Study the online detection technology of trace dissolved oxygen using a three-electrode balance.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Liang Yin.

Ethics declarations

Conflict of interest

The authors declare that they are unaware of competition for financial interests or personal relationships that may affect the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Juan, Y., Ke, Z., Chen, Z. et al. Rapid density estimation of tiny pests from sticky traps using Qpest RCNN in conjunction with UWB-UAV-based IoT framework. Neural Comput & Applic 36, 9779–9803 (2024). https://doi.org/10.1007/s00521-023-09230-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-023-09230-4

Keywords

Navigation