Skip to main content

Graph Convolution and Self Attention Based Non-maximum Suppression

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 11554))

Abstract

Non-maximum suppression is an integral and last part of object detection. Traditional NMS algorithm sorts the detection boxes according to their class scores. The detection boxes with maximum score are always selected while all other boxes with a sufficient overlap with the preserved boxes are discarded. This strategy is simple and effective. However, there still need some improvements in this process because the algorithm makes a ‘hard’ decision (accept or reject) for each box. In this paper, we formulate the non-maximum suppression as a rescoring process and construct a network called NmsNet which utilizes graph convolution and self attention mechanism to predict each box as an object or redundant one. We evaluate our method on the VOC2007 dataset. The experimental results show that our method achieves a higher MAP compared with the traditional greedy NMS and the Soft NMS.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Uijlings, J.R., Van De Sande, K.E., Gevers, T., Smeulders, A.W.: Selective search for object recognition. Int. J. Comput. Vis. 104(2), 154–171 (2013)

    Google Scholar 

  2. Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39(6), 1137–1149 (2017)

    Google Scholar 

  3. Liu, W., et al.: SSD: single shot MultiBox detector. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 21–37. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46448-0_2

    Google Scholar 

  4. Bodla, N., Singh, B., Chellappa, R., Davis, L.S.: Soft-NMS—improving object detection with one line of code. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5562–5570. IEEE Press, Venice (2017)

    Google Scholar 

  5. Hosang, J., Benenson, R., Schiele, B.: learning non-maximum suppression. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6469–6477. IEEE Press, Honolulu (2017)

    Google Scholar 

  6. Ba, J.L., Kiros, J.R., Hinton, G.E.: Layer normalization. arXiv preprint arXiv:1607.06450 (2016)

  7. Niepert, M., Ahmed, M., Kutzkov, K.: Learning convolutional neural networks for graphs. In: Proceedings of the 33rd International Conference on Machine Learning, pp. 2014–2023. ACM, New York (2016)

    Google Scholar 

  8. Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998–6008 (2017)

    Google Scholar 

  9. Yu, A.W., et al.: QANet: combining local convolution with global self-attention for reading comprehension. In: International Conference on Learning Representations (2018)

    Google Scholar 

Download references

Acknowledgement

This work was supported in part by National Natural Science Foundation of China under grants 61771145 and 61371148.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaodong Gu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Qiu, Z., Gu, X. (2019). Graph Convolution and Self Attention Based Non-maximum Suppression. In: Lu, H., Tang, H., Wang, Z. (eds) Advances in Neural Networks – ISNN 2019. ISNN 2019. Lecture Notes in Computer Science(), vol 11554. Springer, Cham. https://doi.org/10.1007/978-3-030-22796-8_9

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-22796-8_9

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-22795-1

  • Online ISBN: 978-3-030-22796-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics