Skip to main content

Advertisement

Log in

Camouflaged object detection based on context-aware and boundary refinement

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Camouflaged Object Detection (COD) has been increasingly studied and the detection performance has been greatly improved based on deep learning models in recent years. However, the context and boundary information have not been efficiently used simultaneously in the existing COD methods, leading to inferior detection for large camouflaged objects, occluded objects, multiple and small objects, and objects with rich boundaries. Therefore, to effectively enhance the performance of COD, we propose a novel camouflaged object detection model, i.e., context-aware and boundary refinement (CABR). Specifically, CABR mainly consists of three modules: the global context information enhanced module (GCIEM), the attention-inducing neighbor fusion module (AINFM), and the boundary refinement module (BRM). GCIEM is designed to fully capture the long-range dependencies to obtain rich global context information to completely detect large objects and occluded objects. AINFM is capable of adaptively fusing adjacent layers to focus on the global and local context information simultaneously to improve the detection performance of multiple and small camouflaged objects effectively. BRM can refine the boundaries by utilizing the spatial information in low-level features and suppressing the non-camouflage factors to detect camouflaged objects with rich boundaries effectively. Quantitative and qualitative experiments are conducted on four benchmark datasets, and the experimental results demonstrate the effectiveness of our CABR with competitive performance to existing state-of-the-art methods according to most evaluation metrics.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9

Similar content being viewed by others

Explore related subjects

Discover the latest articles and news from researchers in related subjects, suggested using machine learning.

Availability of data and materials

The data that support the findings of this study are available on request from the corresponding author.

Code availability

Code is available on request to the authors.

References

  1. Ren Q, Lu S, Zhang J, Hu R (2021) Salient object detection by fusing local and global contexts. IEEE Trans Multimed 23:1442–1453. https://doi.org/10.1109/TMM.2020.2997178

    Article  Google Scholar 

  2. Tu Z, Ma Y, Li C, Tang J, Luo B (2021) Edge-guided non-local fully convolutional network for salient object detection. IEEE Trans Circ Syst Video Technol 31(2):582–593. https://doi.org/10.1109/TCSVT.2020.2980853

    Article  Google Scholar 

  3. Qiu H, Li H, Wu Q, Meng F, Xu L, Ngan KN, Shi H (2020) Hierarchical context features embedding for object detection. IEEE Trans Multimed 22(12):3039–3050. https://doi.org/10.1109/TMM.2020.2971175

    Article  Google Scholar 

  4. Wu Z, Li S, Chen C, Hao A, Qin H (2022) Deeper look at image salient object detection: Bi-stream network with a small training dataset. IEEE Trans Multimed 24:73–86. https://doi.org/10.1109/TMM.2020.3046871

    Article  Google Scholar 

  5. Le TN, Nguyen TV, Nie Z, Tran MT, Sugimoto A (2019) Anabranch network for camouflaged object segmentation. Comput Vis Image Underst 184:45–56

    Article  Google Scholar 

  6. Fan DP, Ji GP, Sun G, Cheng MM, Shen J, Shao L (2020) Camouflaged object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2777–2787

  7. Liu S, Huang D, et al (2018) Receptive field block net for accurate and fast object detection. In: Proceedings of the European conference on computer vision (ECCV), pp 385–400

  8. Mei H, Ji GP, Wei Z, Yang X, Wei X, Fan DP (2021) Camouflaged object segmentation with distraction mining. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 8772–8781

  9. Sun Y, Chen G, Zhou T, Zhang Y, Liu N (2021) Context-aware cross-level fusion network for camouflaged object detection. In: Proceedings of the International Joint Conference on Artificial Intelligence pp 1025–1031

  10. Zhuge M, Lu X, Guo Y, Cai Z, Chen S (2022) Cubenet: X-shape connection for camouflaged object detection. Pattern Recognit 127:108644

    Article  Google Scholar 

  11. Lv Y, Zhang J, Dai Y, Li A, Liu B, Barnes N, Fan DP (2021) Simultaneously localize, segment and rank the camouflaged objects. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 11591–11601

  12. Fan DP, Ji GP, Cheng MM, Shao L (2021) Concealed object detection. IEEE Trans Pattern Anal Mach Intell

  13. Webster RJ (2015) Does disruptive camouflage conceal edges and features? Curr Zool 61(4):708–717

    Article  MathSciNet  Google Scholar 

  14. Webster RJ, Hassall C, Herdman CM, Godin JGJ, Sherratt TN (2013) Disruptive camouflage impairs object recognition. Biol Lett 9(6):20130501

    Article  Google Scholar 

  15. Cao Y, Xu J, Lin S, Wei F, Hu H (2019) Gcnet: Non-local networks meet squeeze-excitation networks and beyond. In: Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, pp 0

  16. Dai Y, Gieseke F, Oehmcke S, Wu Y, Barnard K (2021) Attentional feature fusion. In: Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp 3560–3569

  17. Woo S, Park J, Lee JY, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV), pp 3–19

  18. Skurowski P, Abdulameer H, Błaszczyk J, Depta T, Kornacki A, Kozieł P (2018) Animal camouflage analysis: Chameleon database. Unpublished Manuscr 2(6):7

  19. Bhajantri NU, Nagabhushan P (2006) Camouflage defect identification: a novel approach. In: 9th International Conference on Information Technology (ICIT’06), IEEE, pp 145–148

  20. Boot WR, Neider MB, Kramer AF (2009) Training and transfer of training in the search for camouflaged targets. Atten Percept Psychophys 71(4):950–963

    Article  Google Scholar 

  21. Zhang X, Zhu C, Wang S, Liu Y, Ye M (2016) A bayesian approach to camouflaged moving object detection. IEEE Trans Circ Syst Video Technol 27(9):2001–2013

    Article  Google Scholar 

  22. Pei J, Cheng T, Fan DP, Tang H, Chen C, Van Gool L (2022) Osformer: One-stage camouflaged instance segmentation with transformers. In: Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XVIII, Springer, pp 19–37

  23. Liu Z, Zhang Z, Tan Y, Wu W (2022) Boosting camouflaged object detection with dual-task interactive transformer. In: 2022 26th International Conference on Pattern Recognition (ICPR), IEEE, pp 140–146

  24. Chen S, Tan X, Wang B, Hu X (2018) Reverse attention for salient object detection. In: Proceedings of the European conference on computer vision (ECCV), pp 234–250

  25. Yang F, Zhai Q, Li X, Huang R, Luo A, Cheng H, Fan DP (2021) Uncertainty-guided transformer reasoning for camouflaged object detection. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 4146–4155

  26. Qin X, Fan DP, Huang C, Diagne C, Zhang Z, Sant’Anna AC, Suarez A, Jagersand M, Shao L (2021) Boundary-aware segmentation network for mobile and web applications. arXiv preprint arXiv:2101.04704

  27. Ji GP, Zhu L, Zhuge M, Fu K (2022) Fast camouflaged object detection via edge-based reversible re-calibration network. Pattern Recog 123:108414

    Article  Google Scholar 

  28. Zhou T, Zhou Y, Gong C, Yang J, Zhang Y (2022) Feature aggregation and propagation network for camouflaged object detection. IEEE Trans Image Process 31:7036–7047

    Article  Google Scholar 

  29. Zhai Q, Li X, Yang F, Chen C, Cheng H, Fan DP (2021) Mutual graph learning for camouflaged object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 12997–13007

  30. Ji GP, Fan DP, Chou YC, Dai D, Liniger A, Van Gool L (2022) Deep gradient learning for efficient camouflaged object detection. arXiv preprint arXiv:2205.12853

  31. Zhu J, Zhang X, Zhang S, Liu J (2021) Inferring camouflaged objects by texture-aware interactive guidance network. In: Proceedings of the AAAI Conference on Artificial Intelligence vol 35. pp 3599–3607

  32. Ren J, Hu X, Zhu L, Xu X, Xu Y, Wang W, Deng Z, Heng PA (2021) Deep texture-aware features for camouflaged object detection. IEEE Trans Circ Syst Video Technol

  33. Zhong Y, Li B, Tang L, Kuang S, Wu S, Ding S (2022) Detecting camouflaged object in frequency domain. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 4504–4513

  34. Zhang J, Lv Y, Xiang M, Li A, Dai Y, Zhong Y (2021) Depth confidence-aware camouflaged object detection. arXiv e-prints, 2106

  35. Pang Y, Zhao X, Xiang TZ, Zhang L, Lu H (2022) Zoom in and out: A mixed-scale triplet network for camouflaged object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 2160–2170

  36. Jia Q, Yao S, Liu Y, Fan X, Liu R, Luo Z (2022) Segment, magnify and reiterate: Detecting camouflaged objects the hard way. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 4713–4722

  37. Gao P, Lu K, Xue J, Shao L, Lyu J (2021) A coarse-to-fine facial landmark detection method based on self-attention mechanism. IEEE Trans Multimed 23:926–938. https://doi.org/10.1109/TMM.2020.2991507

    Article  Google Scholar 

  38. Ma X, Guo J, Sansom A, McGuire M, Kalaani A, Chen Q, Tang S, Yang Q, Fu S (2021) Spatial pyramid attention for deep convolutional neural networks. IEEE Trans Multimed 23:3048–3058. https://doi.org/10.1109/TMM.2021.3068576

    Article  Google Scholar 

  39. Yu L, Zhang J, Wu Q (2022) Dual attention on pyramid feature maps for image captioning. IEEE Trans Multimed 24:1775–1786. https://doi.org/10.1109/TMM.2021.3072479

    Article  Google Scholar 

  40. Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132–7141

  41. Huang Z, Wang X, Huang L, Huang C, Wei Y, Liu W (2019) Ccnet: Criss-cross attention for semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 603–612

  42. Tao A, Sapra K, Catanzaro B (2020) Hierarchical multi-scale attention for semantic segmentation. arXiv preprint arXiv:2005.10821

  43. Qin X, Zhang Z, Huang C, Gao C, Dehghan M, Jagersand M (2019) Basnet: Boundary-aware salient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 7479–7489

  44. Feng M, Lu H, Ding E (2019) Attentive feedback network for boundary-aware salient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1623–1632

  45. Liu JJ, Hou Q, Cheng MM, Feng J, Jiang J (2019) A simple pooling-based design for real-time salient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3917–3926

  46. Zhao JX, Liu JJ, Fan DP, Cao Y, Yang J, Cheng MM (2019) Egnet: Edge guidance network for salient object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 8779–8788

  47. Wu Z, Su L, Huang Q (2019) Cascaded partial decoder for fast and accurate salient object detection. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3907–3916

  48. Gao SH, Cheng MM, Zhao K, Zhang XY, Yang MH, Torr P (2019) Res2net: A new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell 43(2):652–662

    Article  Google Scholar 

  49. Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7794–7803

  50. Nair V, Hinton GE (2010) Rectified linear units improve restricted boltzmann machines. In: ICML

  51. Wei J, Wang S, Huang Q (2020) F\(^3\)net: fusion, feedback and focus for salient object detection. In: Proceedings of the AAAI Conference on Artificial Intelligence vol 34. pp 12321–12328

  52. Li W, Zhang Z, Wang X, Luo P (2020) Adax: Adaptive gradient descent with exponential long term memory. arXiv preprint arXiv:2004.09740

  53. Cheng MM, Fan DP (2021) Structure-measure: A new way to evaluate foreground maps. Int J Comput Vis 129(9):2622–2638

    Article  Google Scholar 

  54. Fan DP, Gong C, Cao Y, Ren B, Cheng MM, Borji A (2018) Enhanced-alignment measure for binary foreground map evaluation. In: Proceedings of the International Joint Conference on Artificial Intelligence, pp 698–704

  55. Fan DP, Ji GP, Qin X, Cheng MM (2021) Cognitive vision inspired object segmentation metric and loss function. Sci Sin Inf 6:6

    Google Scholar 

  56. Margolin R, Zelnik-Manor L, Tal A (2014) How to evaluate foreground maps? In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 248–255

  57. Perazzi F, Krähenbühl P, Pritch Y, Hornung A (2012) Saliency filters: Contrast based filtering for salient region detection. In: 2012 IEEE conference on computer vision and pattern recognition, IEEE, pp 733–740

  58. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778

  59. Wu Z, Su L, Huang Q (2019) Stacked cross refinement network for edge-aware salient object detection. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 7264–7273

  60. Zhang J, Fan DP, Dai Y, Anwar S, Saleh FS, Zhang T, Barnes N (2020) Uc-net: Uncertainty inspired rgb-d saliency detection via conditional variational autoencoders. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8582–8591

  61. Li A, Zhang J, Lv Y, Liu B, Zhang T, Dai Y (2021) Uncertainty-aware joint salient object and camouflaged object detection. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp 10071–10081

  62. Yuan Y, Huang L, Guo J, Zhang C, Chen X, Wang J (2018) Ocnet: Object context network for scene parsing. arXiv preprint arXiv:1809.00916

  63. Zhu Z, Xu M, Bai S, Huang T, Bai X (2019) Asymmetric non-local neural networks for semantic segmentation. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp 593–602

  64. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, et al (2020) An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929

  65. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part I 16, pp 213–229

  66. Heo J, Lee HB, Kim S, Lee J, Kim KJ, Yang E, Hwang SJ (2018) Uncertainty-aware attention for reliable interpretation and prediction. Adv Neural Inf Process Syst 31

  67. Yan P, Wu Z, Liu M, Zeng K, Lin L, Li G (2022) Unsupervised domain adaptive salient object detection through uncertainty-aware pseudo-label learning. Proceedings of the AAAI Conference on Artificial Intelligence 36:3000–3008

    Article  Google Scholar 

Download references

Acknowledgements

The authors would like to thank all experts of computer vision for their contribution to camouflaged object detection.

Funding

This work was supported partly by Talent Foundation of Tangshan (A202110011), Distinguished Youth Foundation of North China University of Science and Technology (JQ201715).

Author information

Authors and Affiliations

Authors

Contributions

Caijuan Shi: Conceptualization, Writing- Reviewing and Editing. Bijuan Ren: Conceptualization, Formal analysis, Investigation, Writing - original draft. Houru Chen: Methodology, Programming. Lin Zhao: Visualization, Formal analysis. Chunyu Lin: Writing- Reviewing and Editing. Yao Zhao: Supervision. The author(s) read and approved the final manuscript.

Corresponding author

Correspondence to Caijuan Shi.

Ethics declarations

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Shi, C., Ren, B., Chen, H. et al. Camouflaged object detection based on context-aware and boundary refinement. Appl Intell 53, 22429–22445 (2023). https://doi.org/10.1007/s10489-023-04645-x

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-04645-x

Keywords

Profiles

  1. Chunyu Lin