Skip to main content

Neural Network Verification Accelerated by a Novel Abstract Framework

  • Conference paper
  • First Online:
Advanced Intelligent Computing Technology and Applications (ICIC 2024)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 14864))

Included in the following conference series:

  • 654 Accesses

Abstract

Due to the gradual increase in the application of neural networks in various aspects, ensuring that they do not make errors in critical areas to maintain stability has become a hotspot. The research challenge lies in their black box nature. To address this issue, researchers have proposed various methods for neural network verification. However, the scalability of these methods is limited, making it difficult to apply them to large-scale networks. To address this limitation, this paper proposes an abstract framework. Abstraction, although a classic tool for verification, is rarely used to verify neural networks. However, abstraction aids in addressing the difficulty of extending existing algorithms to the latest network architectures. This framework is applicable to fully connected feedforward neural networks, built upon clustering neurons that exhibit similar output behaviors under specific inputs. For ReLU neural networks, the error bounds generated by the abstraction are also analyzed. Moreover, the framework is shown to minimize the network size while preserving accuracy to the greatest extent possible. We also show how to apply the validation results on the abstract network to the original network. Finally, this framework is independent of existing verification techniques, allowing integration with various verification methods.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Phan-Minh, T, Grigore, E.C., Boulton, F.A., et al.: CoverNet: multimodal behavior prediction using trajectory sets. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 14074–14083 (2020)

    Google Scholar 

  2. Shen, D., Wu, G., Suk, H.I.: Deep learning in medical image analysis. Ann. Rev. Biomed. Eng. 19, 221–248 (2017)

    Google Scholar 

  3. Chang, X., Maekaku, T., Fujita, Y., et al.: End-to-end integration of speech recognition, speech enhancement, and self-supervised learning representation. arXiv preprint arXiv:2204.00540 (2022)

  4. Fischetti, M., Jo, J.: Deep neural networks and mixed integer linear optimization. Constraints 23(3), 296–309 (2018)

    Article  MathSciNet  Google Scholar 

  5. Bunel, R.R., Turkaslan, I., Torr, P., et al.: A unified view of piecewise linear neural network verification. In: Advances in Neural Information Processing Systems, vol. 31 (2018)

    Google Scholar 

  6. Ehlers, R.: Formal verification of piece-wise linear feed-forward neural networks. In: D’Souza, D., Narayan Kumar, K. (eds.) Automated Technology for Verification and Analysis. LNCS, vol. 10482, pp. 269–286. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-68167-2_19

    Chapter  Google Scholar 

  7. Singh, G., Gehr, T., Mirman, M., et al.: Fast and effective robustness certification. In: Advances in Neural Information Processing Systems, vol. 31 (2018)

    Google Scholar 

  8. Singh, G., Gehr, T., Püschel, M., et al.: An abstract domain for certifying neural networks. Proc. ACM Programm. Lang. 3(POPL), 1–30 (2019)

    Google Scholar 

  9. Xiang, W., Tran, H.D., Johnson, T.T.: Output reachable set estimation and verification for multilayer neural networks. IEEE Trans. Neural Netw. Learn. Syst. 29(11), 5777–5783 (2018)

    Article  MathSciNet  Google Scholar 

  10. Wong, E., Kolter, Z.: Provable defenses against adversarial examples via the convex outer adversarial polytope. In: International Conference on Machine Learning, pp. 5286–5295. PMLR (2018)

    Google Scholar 

  11. Clarke, E., Grumberg, O., Jha, S., Lu, Y., Veith, H.: Counterexample-guided abstraction refinement. In: Emerson, E.A., Sistla, A.P. (eds.) Computer Aided Verification. LNCS, vol. 1855, pp. 154–169. Springer, Heidelberg (2000). https://doi.org/10.1007/10722167_15

    Chapter  Google Scholar 

  12. Ashok, P., Hashemi, V., Křetínský, J., Mohr, S.: DeepAbstract: neural network abstraction for accelerating verification. In: Hung, D.V., Sokolsky, O. (eds.) Automated Technology for Verification and Analysis. LNCS, vol. 12302, pp. 92–107. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-59152-6_5

    Chapter  Google Scholar 

Download references

Acknowledgments

This research is funded by Science Research Project of Hebei Education Department under grant No. BJK2024095.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Meng Wang .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Liu, J., Wang, M. (2024). Neural Network Verification Accelerated by a Novel Abstract Framework. In: Huang, DS., Si, Z., Pan, Y. (eds) Advanced Intelligent Computing Technology and Applications. ICIC 2024. Lecture Notes in Computer Science, vol 14864. Springer, Singapore. https://doi.org/10.1007/978-981-97-5588-2_42

Download citation

  • DOI: https://doi.org/10.1007/978-981-97-5588-2_42

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-97-5587-5

  • Online ISBN: 978-981-97-5588-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics