Skip to main content

TPS-YOLO: The Efficient Tiny Person Detection Network Based on Improved YOLOv8 and Model Pruning

  • Conference paper
  • First Online:
MultiMedia Modeling (MMM 2025)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 15523))

Included in the following conference series:

  • 362 Accesses

Abstract

Tiny Person detection in long-range scenes is a popular and challenging task. Current person detectors have two major issues. Firstly, their performance is poor in the case of tiny and heavily occluded persons. Secondly, they are computation-intensive and have large model sizes, which make them difficult to deploy on resource-limited devices. To solve the above issues, we proposed TPS-YOLO. Based on YOLOv8, we reconstruct the network structure by introducing shallow features of P2 into the feature fusion layers, which helps retain more spatial information important for tiny person detection. We design a fine-grained feature extraction module SPDCA to replace the standard convolution layer in the backbone network to enhance the feature representation of the network. In the feature fusion network, we use a weighted fusion method to fuse multi-scale features, which introduces learnable weights to learn the importance of different input features. We propose a lightweight module named C2f_Efficient, which integrates Depthwise Separable Convolution (DSC) to reduce the model parameters. Furthermore, we apply a model pruning method to further reduce the model’s computational complexity. Experiments on the Tinypersonv2 and VisDrone-person datasets show that TPS-YOLO achieves satisfactory performance in terms of both efficiency and accuracy and has advantages on model lightweight.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Jiang, N., Yu, X., Peng, X., Gong, Y., Han, Z.: SM+: refined scale match for tiny person detection (2021)

    Google Scholar 

  2. Peng, G., Yang, Z., Wang, S., Zhou, Y.: AMFLW-YOLO: a lightweight network for remote sensing image detection based on attention mechanism and multi-scale feature fusion. IEEE Trans. Geosci. Remote Sens. 16 (2023)

    Google Scholar 

  3. Cao, J., Pang, Y., Xie, J., Khan, F.S., Shao, L.: From handcrafted to deep features for pedestrian detection: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 44(9) 4913–4934 (2021)

    Google Scholar 

  4. Khan, A.H., Nawaz, M.S., Dengel, A.: Localized semantic feature mixers for efficient pedestrian detection in autonomous driving. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5476–5485 (2023)

    Google Scholar 

  5. Redmon, J., Divvala, S., Girshick, R., Farhadi. A.: You only look once: unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779–788 (2016)

    Google Scholar 

  6. Redmon .J., Farhadi. A.: YOLOv3: an incremental improvement. arXiv preprint

    Google Scholar 

  7. Wang, C.Y., Yeh, I.H., Liao, H.Y.M.: YOLOV9: learning what you want to learn using programmable gradient information. arXiv preprintarXiv:2402.13616 (2024)

    Google Scholar 

  8. Shi, Y., Li, S., Liu, Z., Zhou, Z., Zhou, X.: MTP-YOLO: you only look once based maritime tiny person detector for emergency rescue. J. Marine Sci. Eng. 12(4) (2024)

    Google Scholar 

  9. Lin, T.-Y., Dollár, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117–2125 (2017)

    Google Scholar 

  10. Li, G., Yang, Y., Xingda, Q.: Deep learning approaches on pedestrian detection in hazy weather. IEEE Trans. Industr. Electron. 67(10), 8889–8899 (2019)

    Article  MATH  Google Scholar 

  11. Howard, A., et al.: Searching for MobileNetV3. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 1314–1324 (2019)

    Google Scholar 

  12. Kim, B.J., Choi, H., Jang, H., Lee, D.G., Jeong, W., Kim, S.W.: Dead pixel test using effective receptive field. Pattern Recogn. Lett, 167, 149–156 (2023)

    Google Scholar 

  13. Sunkara, R., Luo, T.: No more strided convolutions or pooling: a new CNN building block for low-resolution images and small objects. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 443–459. Springer (2022)

    Google Scholar 

  14. Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 13713–13722 (2021)

    Google Scholar 

  15. Hua, B.-S., Tran, M.-K., Yeung, S.-K.: Pointwise convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 984–993 (2018)

    Google Scholar 

  16. Tan, M., Pang, R., Le, Q.V.: EfficientDet: scalable and efficient object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10781–10790 (2020)

    Google Scholar 

  17. Zhang, X., Zhou, X., Lin, M., Sun, J.: ShuffleNet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848–6856 (2018)

    Google Scholar 

  18. Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251–1258 (2017)

    Google Scholar 

  19. Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132–7141 (2018)

    Google Scholar 

  20. Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., Zhang, C.: Learning efficient convolutional networks through network slimming. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2736–2744 (2017)

    Google Scholar 

  21. Yu, X., et al.: Object localization under single coarse point supervision. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4868–4877 (2022)

    Google Scholar 

  22. Zhu, P., Wen, L., Bian, X., Ling, H., Hu, Q.: Vision meets drones: A challenge. arXiv preprint arXiv:1804.07437 (2018)

  23. Yu, X., Gong, Y., Jiang, N., Ye, Q., Han, Z.: Scale match for tiny person detection. In: Workshop on Applications of Computer Vision (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Qianni Huang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2025 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yao, L., Huang, Q., Wan, Y. (2025). TPS-YOLO: The Efficient Tiny Person Detection Network Based on Improved YOLOv8 and Model Pruning. In: Ide, I., et al. MultiMedia Modeling. MMM 2025. Lecture Notes in Computer Science, vol 15523. Springer, Singapore. https://doi.org/10.1007/978-981-96-2071-5_18

Download citation

  • DOI: https://doi.org/10.1007/978-981-96-2071-5_18

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-96-2070-8

  • Online ISBN: 978-981-96-2071-5

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics