Skip to main content

A Bottom-Up Approach to Applying Graphical Models in Security Analysis

  • Conference paper
  • First Online:
Graphical Models for Security (GraMSec 2016)

Part of the book series: Lecture Notes in Computer Science ((LNSC,volume 9987))

Included in the following conference series:

  • 690 Accesses

Abstract

Graphical models have emerged as a widely adopted approach to conducting security analysis for computer and network systems. The power of graphical models lies in two aspects: the graph structure can be used to capture correlations among security events, and the quantitative reasoning over the graph structure can render useful triaging decisions when dealing with the inherent uncertainty in security events. In this work we leverage these powers afforded by graphical model in security analysis. Given that the analyst is the intended user of the model, the most difficult task for research in this area is to understand the real world constraints under which security analysts must operate with. Those constraints dictate what parameters are realistically obtainable to use in the designed graphical models, and what type of reasoning results can be useful to analysts. We present how we use this bottom-up approach to design customized graphical models for enterprise network intrusion analysis. In this work, we had to design specific graph generation algorithms based on the concrete security problems at hands, and customized reasoning algorithms to use the graphical model to yield useful tools for analysts.

This article was based on a previously published work [38].

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Notes

  1. 1.

    Only training data’s truth file is publicly available.

References

  1. Snort rules documentation. http://www.snort.org

  2. Almgren, M., Lindqvist, U., Jonsson, E.: A multi-sensor model to improve automated attack detection. In: 11th International Symposium on Recent Advances in Intrusion Detection (RAID 2008). RAID, September 2008

    Google Scholar 

  3. Axelsson, S.: The base-rate fallacy and the difficulty of intrusion detection. ACM Trans. Inf. Syst. Secur. 3(3), 186–205 (2000)

    Article  MathSciNet  Google Scholar 

  4. Barreno, M., Cárdenas, A.A., Tygar, J.D.: Optimal ROC curve for a combination of classifiers. In: Advances in Neural Information Processing Systems (NIPS, 2007) (2008)

    Google Scholar 

  5. Carrier, B.: A hypothesis-based approach to digital forensic investigations. Technical report, Center for Education and Research in Information Assurance and Security (CERIAS), Purdue University (2006)

    Google Scholar 

  6. Chen, Q., Aickelin, U.: Anomaly detection using the Dempster-Shafer method. In: International Conference on Data Mining (DMIN 2006) (2006)

    Google Scholar 

  7. Chen, T.M., Venkataramanan, V.: Dempster-Shafer theory for intrusion detection in ad hoc networks. IEEE Internet Comput. 9, 35–41 (2005)

    Article  Google Scholar 

  8. Cheung, S., Lindqvist, U., Fong, M.W.: Modeling multistep cyber attacks for scenario recognition. In: DARPA Information Survivability Conference and Exposition (DISCEX III), Washington, D.C., pp. 284–292 (2003)

    Google Scholar 

  9. Cuppens, F., Miège, A.: Alert correlation in a cooperative intrusion detection framework. In: IEEE Symposium on Security and Privacy (2002)

    Google Scholar 

  10. Denceux, T.: The cautious rule of combination for belief functions and some extensions. In: 9th International Conference on Information Fusion (2006)

    Google Scholar 

  11. Fine, T.L.: Theories of Probability. Academic Press, New York (1973)

    MATH  Google Scholar 

  12. Guofei, G., Cárdenas, A.A., Lee, W.: Principled reasoning and practical applications of alert fusion in intrusion detection systems. In: Proceedings of the 2008 ACM Symposium on Information, Computer and Communications Security, ASIACCS 2008, pp. 136–147. ACM, New York (2008)

    Google Scholar 

  13. Halpern, J.Y.: Reasoning About Uncertainty. The MIT Press, London (2005)

    MATH  Google Scholar 

  14. Jensen, F.V., Nielsen, T.D.: Bayesian Networks and Decision Graphs. Springer, New York (2007)

    Google Scholar 

  15. ArgusLab.: Snort intrusion analysis using proof strengthening (SnIPS). http://people.cis.ksu.edu/xou/argus/software/snips/

  16. Mahoney, M.V., Chan, P.K.: An analysis of the 1999 DARPA/Lincoln laboratory evaluation data for network anomaly detection. In: Proceedings of the Sixth International Symposium on Recent Advances in Intrusion Detection (RAID) (2003)

    Google Scholar 

  17. McHugh, J.: Testing intrusion detection systems: a critique of the 1998 and 1999 DARPA intrusion detection system evaluations as performed by Lincoln laboratory. ACM Trans. Inf. Syst. Secur. (TISSEC) 3(4), 262–294 (2000)

    Article  Google Scholar 

  18. Modelo-Howard, G., Bagchi, S., Lebanon, G.: Determining placement of intrusion detectors for a distributed application through Bayesian network modeling. In: 11th International Symposium on Recent Advances in Intrusion Detection (RAID 2008). RAID, September 2008

    Google Scholar 

  19. Ning, P., Cui, Y., Reeves, D., Dingbang, X.: Tools and techniques for analyzing intrusion alerts. ACM Trans. Inf. Syst. Secur. 7(2), 273–318 (2004)

    Article  Google Scholar 

  20. Noel, S., Robertson, E., Jajodia, S.: Correlating intrusion events and building attack scenarios through attack graph distances. In: 20th Annual Computer Security Applications Conference (ACSAC 2004), pp. 350–359 (2004)

    Google Scholar 

  21. Xinming, O., Raj Rajagopalan, S., Sakthivelmurugan, S.: An empirical approach to modeling uncertainty in intrusion analysis. In: Annual Computer Security Applications Conference (ACSAC), December 2009

    Google Scholar 

  22. Sentz, K., Ferson, S.: Combination of evidence in Dempster-Shafer theory. Technical report, Sandia National Laboratories, Albuquerque, New Mexico (2002)

    Google Scholar 

  23. Shafer, G.: A Mathematical Theory of Evidence. Princeton University Press, Princeton (1976)

    MATH  Google Scholar 

  24. Shafer, G.: The problem of dependent evidence. Technical report, University of Kansas (1984)

    Google Scholar 

  25. Shafer, G.: Belief functions and possibility measures. In: The Analysis of Fuzzy Information (1986)

    Google Scholar 

  26. Shafer, G.: Probability judgment in artificial intelligence and expert systems. Stat. Sci. 2(1), 3–16 (1987)

    Article  MathSciNet  MATH  Google Scholar 

  27. Smith, R., Japkowicz, N., Dondo, M., Mason, P.: Using unsupervised learning for network alert correlation. In: Bergler, S. (ed.) Canadian AI. LNCS (LNAI), vol. 5032, pp. 308–319. Springer, Heidelberg (2008)

    Chapter  Google Scholar 

  28. Sun, L., Srivastava, R.P., Mock, T.J.: An information systems security risk assessment model under Dempster-Shafer theory of belief functions. J. Manag. Inf. 22, 109–142 (2006)

    Google Scholar 

  29. Sundaramurthy, S.C., Zomlot, L., Xinming, O.: Practical IDS alert correlation in the face of dynamic threats. In: The 2011 International Conference on Security and Management (SAM 2011), Las Vegas, USA, July 2011

    Google Scholar 

  30. Svensson, H., Audun J\(\phi \)sang.: Correlation of intrusion alarms with subjective logic. In: Sixth Nordic Workshop on Secure IT systems (NordSec) (2001)

    Google Scholar 

  31. Valeur, F.: Real-time intrusion detection alert correlation. Ph.D. thesis, University of California, Santa Barbara, May 2006

    Google Scholar 

  32. Valeur, F., Vigna, G., Kruegel, C., Kemmerer, R.A.: A comprehensive approach to intrusion detection alert correlation. IEEE Trans. Dependable Secure Comput. 1(3), 146–169 (2004)

    Article  Google Scholar 

  33. Dong, Y., Frincke, D.: A novel framework for alert correlation and understanding. In: International Conference on Applied Cryptography and Network Security (ACNS) (2004)

    Google Scholar 

  34. Dong, Y., Frincke, D.: Alert confidence fusion in intrusion detection systems with extended Dempster-Shafer theory. In: 43rd ACM Southeast Conference, Kennesaw, GA, USA (2005)

    Google Scholar 

  35. Zhai, Y., Ning, P., Xu, J.: Integrating IDS alert correlation and OS-level dependency tracking. In: Mehrotra, S., Zeng, D.D., Chen, H., Thuraisingham, B., Wang, F.-Y. (eds.) ISI 2006. LNCS, vol. 3975, pp. 272–284. Springer, Heidelberg (2006)

    Chapter  Google Scholar 

  36. Zhai, Y., Ning, P., Iyer, P., Reeves, D.S.: Reasoning about complementary intrusion evidence. In: Proceedings of 20th Annual Computer Security Applications Conference (ACSAC), pp. 39–48, December 2004

    Google Scholar 

  37. Zhou, J., Heckman, M., Reynolds, B., Carlson, A., Bishop, M.: Modeling network intrusion detection alerts for correlation. ACM Trans. Inf. Syst. Secur. (TISSEC) 10(1), 4 (2007)

    Article  Google Scholar 

  38. Zomlot, L., Sundaramurthy, S.C., Luo, K., Xinming, O., Raj Rajagopalan, S.: Prioritizing intrusion analysis using Dempster-Shafer theory. In: 4TH ACM Workshop on Artificial Intelligence and Security (AISec) (2011)

    Google Scholar 

Download references

Acknowledgments

This material is based upon work supported by U.S. National Science Foundation under grant no. 1622402 and 1314925, AFOSR under Award No. FA9550-09-1-0138, and HP Labs Innovation Research Program. Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation, AFOSR, or Hewlett-Packard Development Company, L.P.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xinming Ou .

Editor information

Editors and Affiliations

Appendices

A Semantics of the Overlapping Factors

Since we only have two non-zero bpa subsets: t and \(\theta \), in each hypothesis’s frame of discernment, we use \(w_i\) to denote the fact that we trust \(h_i\) (\(h_i=t\)) and \(\bar{w_i}\) (negation of \(w_i\)) to denote the fact that we do not trust \(h_i\) (\(h_i=\theta \)). One may find it strange that \(w_i\) and \(\bar{w_i}\) appear to be not mutually exclusive, since \(\theta \) includes both t and f. This is exactly the unique way in which DS expresses disbelief in a hypothesis – it differentiates clearly between not believing a hypothesis and believing the negation of that hypothesis. When we trust a hypothesis, we believe its state is t and when we do not trust a hypothesis, we do not know what its state is, hence \(\theta \). Interested readers are referred to Shafer’s discussion on how to handle non-independent evidence using this interpretation [26]. The semantics of overlapping factor can be defined as:

$$\begin{aligned}&r_1 = \frac{Pr[w_2|w_1] - Pr[w_2]}{Pr[\bar{w_{2}}]}, r_2 = \frac{Pr[w_1|w_2] - Pr[w_1]}{Pr[\bar{w_{1}}]} \end{aligned}$$

Let us take \(r_1\) as an example to explain the semantics. If we condition on trusting hypothesis \(h_1\), the probability that we also trust \(h_2\) is greater than or equal to its absolute probability since shared IDS sensors only give us positive correlation. The bigger the difference, the stronger influence trusting \(h_1\) has on trusting \(h_2\). The extreme case is when \(Pr[w_2|w_1] = 1\), which gives \(r_1 = 1\). Both \(r_1\) and \(r_2\) measure the dependence between \(w_1\) and \(w_2\), but from different directions.

Theorem A1

$$\begin{aligned}&r_{2}= \alpha \cdot r_1 , \textit{ where } \alpha =\frac{Pr[w_{1}] \cdot Pr[\bar{w_2}]}{Pr[w_{2}] \cdot Pr[\bar{w_1}]} \end{aligned}$$
(14)

Proof

$$\begin{aligned}&r_1\cdot Pr[\bar{w_2}]\cdot Pr[w_1]=Pr[w_1,w_2]-Pr[w_1]\cdot Pr[w_2]\\&r_2\cdot Pr[\bar{w_1}]\cdot Pr[w_2]=Pr[w_1,w_2]-Pr[w_1]\cdot Pr[w_2] \end{aligned}$$

We then have

$$\begin{aligned}&r_1\cdot Pr[\bar{w_2}]\cdot Pr[w_1]=r_2\cdot Pr[\bar{w_1}]\cdot Pr[w_2] \end{aligned}$$

Theorem A2

$$\begin{aligned} \psi [h_1 , h_2] = Pr[w_1 , w_2] \end{aligned}$$

Proof

Let us substitute \(r_i\)’s into formulas (7)–(10). Let us also substitute the following definitions:

$$\begin{aligned} \begin{array}{ll} m_i(t) = Pr[w_i]&m_i(\theta )= Pr[\bar{w_i}] \end{array} \end{aligned}$$

knowing that:

$$\begin{aligned}&Pr[w_2|w_1]=\frac{Pr[w_1,w_2]}{Pr[w_1]} \end{aligned}$$

then substitute the above into the definition of \(r_1\), we get

$$\begin{aligned}&r_1\cdot Pr[\bar{w_2}]\cdot Pr[w_1]=Pr[w_1,w_2]-Pr[w_1]\cdot Pr[w_2] \end{aligned}$$

knowing that \(Pr[\bar{w_2}]=1-Pr[w_2]\), then:

$$\begin{aligned}&Pr[w_1,w_2]=r_{1} \cdot Pr[w_{1}] +(1-r_{1}) \cdot Pr[w_{1}] \cdot Pr[w_{2}] \\&=\psi [t,t] \end{aligned}$$

The importance of this theorem is that our way of calculating the joint bpa \(\psi [h_1, h_2]\) is sound in that it gives a generalization of the joint probability distribution of the trustworthiness of two (potentially) dependent sources. This also follows Shafer’s general guide on how to handle non-independent evidence sources in DS [26], although Shafer did not provide the specific formulations.

B Belief Calculation Algorithm

The main algorithm is DsCorr (Algorithm 1). This function takes GraphSet which is a set of correlation graph segments. It iterates on each graph, and returns a set of the graph segments sorted by the belief of the sink node (or highest sink node for multiple sinks) in descending order.

figure a

Algorithm ComputeNodeBelief (Algorithm 2) takes a node and returns the belief value of it. There are three cases to consider for the node: (1) it is a source node; (2) it has only one parent node, (3) it has multiple parents. In the first case AssignBpaValues is called to compute the basic probability assignment based on the method in Sect. 3.2. This case applies to the alert nodes, e.g., node 1–5 in Fig. 1. In the second case the node has only one parent so the translation function is called. The third case for combination is done by first translating implicitly into the node and then combine.

figure b

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing AG

About this paper

Cite this paper

Ou, X. (2016). A Bottom-Up Approach to Applying Graphical Models in Security Analysis. In: Kordy, B., Ekstedt, M., Kim, D. (eds) Graphical Models for Security. GraMSec 2016. Lecture Notes in Computer Science(), vol 9987. Springer, Cham. https://doi.org/10.1007/978-3-319-46263-9_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-46263-9_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-46262-2

  • Online ISBN: 978-3-319-46263-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics