skip to main content
10.1145/3647649.3647653acmotherconferencesArticle/Chapter ViewAbstractPublication PagesicigpConference Proceedingsconference-collections
research-article

Contextual Boundary Aware Network for Salient Object Detection

Published:03 May 2024Publication History

ABSTRACT

Currently, for the task of salient object detection (SOD) based on deep learning, most approaches use a strategy of multi-level feature aggregation to enhance performance. However, due to the insufficient utilization of inter-pixel information, the aggregation of multi-level features often affects the prediction of salient objects and results in detecting blurry boundaries of salient objects. To tackle this problem, we have proposed a salient object detection network based on context-aware boundary perception. This network utilizes the context awareness (CA) branch to extract comprehensive contextual semantic information, guiding the network to focus attention not only on salient objects, but also on learning the mutual relationships between multiple salient objects. In addition, the boundary awareness (BA) branch is utilized to explore detailed boundary information around the contours of salient objects, enhancing the network's understanding of the edge pixels of salient objects. Moreover, we have introduced a new feature interaction aggregation (FIA) module, which is used to merge contextual semantic information and boundary detail information in the decoding stage to effectively utilize multi-level features and generate clearer and more accurate saliency maps. By conducting comprehensive experiments on three public datasets, we have demonstrated that our proposed method outperforms the current state-of-the-art representative methods.

References

  1. J. Long, E. Shelhamer, T. Darrell, Fully convolutional networks for semantic segmentation, in: IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 3431–3440 .Google ScholarGoogle ScholarCross RefCross Ref
  2. Zhou, H., **e, X., Lai, J. H., Chen, Z., & Yang, L. (2020). Interactive two-stream decoder for accurate and fast saliency detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 9141-9150).Google ScholarGoogle ScholarCross RefCross Ref
  3. Wei, J., Wang, S., Wu, Z., Su, C., Huang, Q., & Tian, Q. (2020). Label decoupling framework for salient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 13025-13034).Google ScholarGoogle ScholarCross RefCross Ref
  4. Pang, Y., Zhao, X., Zhang, L., & Lu, H. (2020). Multi-scale interactive network for salient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 9413-9422).Google ScholarGoogle ScholarCross RefCross Ref
  5. Cheng, M. M., Mitra, N. J., Huang, X., Torr, P. H., & Hu, S. M. (2014). Global contrast based salient region detection. IEEE transactions on pattern analysis and machine intelligence, 37(3), 569-582.Google ScholarGoogle Scholar
  6. Zhu, W., Liang, S., Wei, Y., & Sun, J. (2014). Saliency optimization from robust background detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 2814-2821).Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Lee, G., Tai, Y. W., & Kim, J. (2016). Deep saliency with encoded low level distance map and high level features. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 660-668).Google ScholarGoogle ScholarCross RefCross Ref
  8. Li, G., & Yu, Y. (2015). Visual saliency based on multiscale deep features. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 5455-5463).Google ScholarGoogle Scholar
  9. Zhao, X., Pang, Y., Zhang, L., Lu, H., & Zhang, L. (2020). Suppress and balance: A simple gated network for salient object detection. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16 (pp. 35-51). Springer International Publishing.Google ScholarGoogle Scholar
  10. Zhang, Q., Shi, Y., Zhang, X., & Zhang, L. (2022). Residual attentive feature learning network for salient object detection. Neurocomputing, 501, 741-752.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Li, J., Pan, Z., Liu, Q., & Wang, Z. (2020). Stacked U-shape network with channel-wise attention for salient object detection. IEEE Transactions on Multimedia, 23, 1397-1409.Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Hui, S., Guo, Q., Geng, X., & Zhang, C. (2023). Multi-Guidance CNNs for Salient Object Detection. ACM Transactions on Multimedia Computing, Communications and Applications, 19(3), 1-19.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. Feng M, Lu H, Ding E. Attentive feedback network for boundary-aware salient object detection[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2019: 1623-1632.Google ScholarGoogle Scholar
  14. Liu, N., Han, J., & Yang, M. H. (2018). Picanet: Learning pixel-wise contextual attention for saliency detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 3089-3098).Google ScholarGoogle ScholarCross RefCross Ref
  15. Wu, Z., Su, L., & Huang, Q. (2019). Cascaded partial decoder for fast and accurate salient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (pp. 3907-3916).Google ScholarGoogle ScholarCross RefCross Ref
  16. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 770-778).Google ScholarGoogle ScholarCross RefCross Ref
  17. Yun Y K, Lin W. Selfreformer: Self-refined network with transformer for salient object detection[J]. arXiv preprint arXiv:2205.11283, 2022.Google ScholarGoogle Scholar
  18. Zhang J, Shi Y, Zhang Q, Attention guided contextual feature fusion network for salient object detection[J]. Image and Vision Computing, 2022, 117: 104337.Google ScholarGoogle Scholar
  19. Qin X, Zhang Z, Huang C, Basnet: Boundary-aware salient object detection[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2019: 7479-7489.Google ScholarGoogle Scholar
  20. Máttyus G, Luo W, Urtasun R. Deeproadmapper: Extracting road topology from aerial images[C]//Proceedings of the IEEE international conference on computer vision. 2017: 3438-3446.Google ScholarGoogle Scholar
  21. Wang L, Lu H, Wang Y, Learning to detect salient objects with image-level supervision[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2017: 136-145.Google ScholarGoogle Scholar
  22. Yan Q, Xu L, Shi J, Hierarchical saliency detection[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2013: 1155-1162.Google ScholarGoogle Scholar
  23. Li G, Yu Y. Visual saliency based on multiscale deep features[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2015: 5455-5463.Google ScholarGoogle Scholar
  24. He K, Zhang X, Ren S, Delving deep into rectifiers: Surpassing human-level performance on imagenet classification[C]//Proceedings of the IEEE international conference on computer vision. 2015: 1026-1034.Google ScholarGoogle Scholar
  25. Margolin R, Zelnik-Manor L, Tal A. How to evaluate foreground maps?[C]//Proceedings of the IEEE conference on computer vision and pattern recognition. 2014: 248-255.Google ScholarGoogle Scholar
  26. Fan D P, Gong C, Cao Y, Enhanced-alignment measure for binary foreground map evaluation[J]. arXiv preprint arXiv:1805.10421, 2018.Google ScholarGoogle Scholar
  27. Fan D P, Cheng M M, Liu Y, Structure-measure: A new way to evaluate foreground maps[C]//Proceedings of the IEEE international conference on computer vision. 2017: 4548-4557.Google ScholarGoogle Scholar

Index Terms

  1. Contextual Boundary Aware Network for Salient Object Detection

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      ICIGP '24: Proceedings of the 2024 7th International Conference on Image and Graphics Processing
      January 2024
      480 pages
      ISBN:9798400716720
      DOI:10.1145/3647649

      Copyright © 2024 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 3 May 2024

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited
    • Article Metrics

      • Downloads (Last 12 months)8
      • Downloads (Last 6 weeks)8

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format