Abstract
Graphical models have emerged as a widely adopted approach to conducting security analysis for computer and network systems. The power of graphical models lies in two aspects: the graph structure can be used to capture correlations among security events, and the quantitative reasoning over the graph structure can render useful triaging decisions when dealing with the inherent uncertainty in security events. In this work we leverage these powers afforded by graphical model in security analysis. Given that the analyst is the intended user of the model, the most difficult task for research in this area is to understand the real world constraints under which security analysts must operate with. Those constraints dictate what parameters are realistically obtainable to use in the designed graphical models, and what type of reasoning results can be useful to analysts. We present how we use this bottom-up approach to design customized graphical models for enterprise network intrusion analysis. In this work, we had to design specific graph generation algorithms based on the concrete security problems at hands, and customized reasoning algorithms to use the graphical model to yield useful tools for analysts.
This article was based on a previously published work [38].
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
Only training data’s truth file is publicly available.
References
Snort rules documentation. http://www.snort.org
Almgren, M., Lindqvist, U., Jonsson, E.: A multi-sensor model to improve automated attack detection. In: 11th International Symposium on Recent Advances in Intrusion Detection (RAID 2008). RAID, September 2008
Axelsson, S.: The base-rate fallacy and the difficulty of intrusion detection. ACM Trans. Inf. Syst. Secur. 3(3), 186–205 (2000)
Barreno, M., Cárdenas, A.A., Tygar, J.D.: Optimal ROC curve for a combination of classifiers. In: Advances in Neural Information Processing Systems (NIPS, 2007) (2008)
Carrier, B.: A hypothesis-based approach to digital forensic investigations. Technical report, Center for Education and Research in Information Assurance and Security (CERIAS), Purdue University (2006)
Chen, Q., Aickelin, U.: Anomaly detection using the Dempster-Shafer method. In: International Conference on Data Mining (DMIN 2006) (2006)
Chen, T.M., Venkataramanan, V.: Dempster-Shafer theory for intrusion detection in ad hoc networks. IEEE Internet Comput. 9, 35–41 (2005)
Cheung, S., Lindqvist, U., Fong, M.W.: Modeling multistep cyber attacks for scenario recognition. In: DARPA Information Survivability Conference and Exposition (DISCEX III), Washington, D.C., pp. 284–292 (2003)
Cuppens, F., Miège, A.: Alert correlation in a cooperative intrusion detection framework. In: IEEE Symposium on Security and Privacy (2002)
Denceux, T.: The cautious rule of combination for belief functions and some extensions. In: 9th International Conference on Information Fusion (2006)
Fine, T.L.: Theories of Probability. Academic Press, New York (1973)
Guofei, G., Cárdenas, A.A., Lee, W.: Principled reasoning and practical applications of alert fusion in intrusion detection systems. In: Proceedings of the 2008 ACM Symposium on Information, Computer and Communications Security, ASIACCS 2008, pp. 136–147. ACM, New York (2008)
Halpern, J.Y.: Reasoning About Uncertainty. The MIT Press, London (2005)
Jensen, F.V., Nielsen, T.D.: Bayesian Networks and Decision Graphs. Springer, New York (2007)
ArgusLab.: Snort intrusion analysis using proof strengthening (SnIPS). http://people.cis.ksu.edu/xou/argus/software/snips/
Mahoney, M.V., Chan, P.K.: An analysis of the 1999 DARPA/Lincoln laboratory evaluation data for network anomaly detection. In: Proceedings of the Sixth International Symposium on Recent Advances in Intrusion Detection (RAID) (2003)
McHugh, J.: Testing intrusion detection systems: a critique of the 1998 and 1999 DARPA intrusion detection system evaluations as performed by Lincoln laboratory. ACM Trans. Inf. Syst. Secur. (TISSEC) 3(4), 262–294 (2000)
Modelo-Howard, G., Bagchi, S., Lebanon, G.: Determining placement of intrusion detectors for a distributed application through Bayesian network modeling. In: 11th International Symposium on Recent Advances in Intrusion Detection (RAID 2008). RAID, September 2008
Ning, P., Cui, Y., Reeves, D., Dingbang, X.: Tools and techniques for analyzing intrusion alerts. ACM Trans. Inf. Syst. Secur. 7(2), 273–318 (2004)
Noel, S., Robertson, E., Jajodia, S.: Correlating intrusion events and building attack scenarios through attack graph distances. In: 20th Annual Computer Security Applications Conference (ACSAC 2004), pp. 350–359 (2004)
Xinming, O., Raj Rajagopalan, S., Sakthivelmurugan, S.: An empirical approach to modeling uncertainty in intrusion analysis. In: Annual Computer Security Applications Conference (ACSAC), December 2009
Sentz, K., Ferson, S.: Combination of evidence in Dempster-Shafer theory. Technical report, Sandia National Laboratories, Albuquerque, New Mexico (2002)
Shafer, G.: A Mathematical Theory of Evidence. Princeton University Press, Princeton (1976)
Shafer, G.: The problem of dependent evidence. Technical report, University of Kansas (1984)
Shafer, G.: Belief functions and possibility measures. In: The Analysis of Fuzzy Information (1986)
Shafer, G.: Probability judgment in artificial intelligence and expert systems. Stat. Sci. 2(1), 3–16 (1987)
Smith, R., Japkowicz, N., Dondo, M., Mason, P.: Using unsupervised learning for network alert correlation. In: Bergler, S. (ed.) Canadian AI. LNCS (LNAI), vol. 5032, pp. 308–319. Springer, Heidelberg (2008)
Sun, L., Srivastava, R.P., Mock, T.J.: An information systems security risk assessment model under Dempster-Shafer theory of belief functions. J. Manag. Inf. 22, 109–142 (2006)
Sundaramurthy, S.C., Zomlot, L., Xinming, O.: Practical IDS alert correlation in the face of dynamic threats. In: The 2011 International Conference on Security and Management (SAM 2011), Las Vegas, USA, July 2011
Svensson, H., Audun J\(\phi \)sang.: Correlation of intrusion alarms with subjective logic. In: Sixth Nordic Workshop on Secure IT systems (NordSec) (2001)
Valeur, F.: Real-time intrusion detection alert correlation. Ph.D. thesis, University of California, Santa Barbara, May 2006
Valeur, F., Vigna, G., Kruegel, C., Kemmerer, R.A.: A comprehensive approach to intrusion detection alert correlation. IEEE Trans. Dependable Secure Comput. 1(3), 146–169 (2004)
Dong, Y., Frincke, D.: A novel framework for alert correlation and understanding. In: International Conference on Applied Cryptography and Network Security (ACNS) (2004)
Dong, Y., Frincke, D.: Alert confidence fusion in intrusion detection systems with extended Dempster-Shafer theory. In: 43rd ACM Southeast Conference, Kennesaw, GA, USA (2005)
Zhai, Y., Ning, P., Xu, J.: Integrating IDS alert correlation and OS-level dependency tracking. In: Mehrotra, S., Zeng, D.D., Chen, H., Thuraisingham, B., Wang, F.-Y. (eds.) ISI 2006. LNCS, vol. 3975, pp. 272–284. Springer, Heidelberg (2006)
Zhai, Y., Ning, P., Iyer, P., Reeves, D.S.: Reasoning about complementary intrusion evidence. In: Proceedings of 20th Annual Computer Security Applications Conference (ACSAC), pp. 39–48, December 2004
Zhou, J., Heckman, M., Reynolds, B., Carlson, A., Bishop, M.: Modeling network intrusion detection alerts for correlation. ACM Trans. Inf. Syst. Secur. (TISSEC) 10(1), 4 (2007)
Zomlot, L., Sundaramurthy, S.C., Luo, K., Xinming, O., Raj Rajagopalan, S.: Prioritizing intrusion analysis using Dempster-Shafer theory. In: 4TH ACM Workshop on Artificial Intelligence and Security (AISec) (2011)
Acknowledgments
This material is based upon work supported by U.S. National Science Foundation under grant no. 1622402 and 1314925, AFOSR under Award No. FA9550-09-1-0138, and HP Labs Innovation Research Program. Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation, AFOSR, or Hewlett-Packard Development Company, L.P.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendices
A Semantics of the Overlapping Factors
Since we only have two non-zero bpa subsets: t and \(\theta \), in each hypothesis’s frame of discernment, we use \(w_i\) to denote the fact that we trust \(h_i\) (\(h_i=t\)) and \(\bar{w_i}\) (negation of \(w_i\)) to denote the fact that we do not trust \(h_i\) (\(h_i=\theta \)). One may find it strange that \(w_i\) and \(\bar{w_i}\) appear to be not mutually exclusive, since \(\theta \) includes both t and f. This is exactly the unique way in which DS expresses disbelief in a hypothesis – it differentiates clearly between not believing a hypothesis and believing the negation of that hypothesis. When we trust a hypothesis, we believe its state is t and when we do not trust a hypothesis, we do not know what its state is, hence \(\theta \). Interested readers are referred to Shafer’s discussion on how to handle non-independent evidence using this interpretation [26]. The semantics of overlapping factor can be defined as:
Let us take \(r_1\) as an example to explain the semantics. If we condition on trusting hypothesis \(h_1\), the probability that we also trust \(h_2\) is greater than or equal to its absolute probability since shared IDS sensors only give us positive correlation. The bigger the difference, the stronger influence trusting \(h_1\) has on trusting \(h_2\). The extreme case is when \(Pr[w_2|w_1] = 1\), which gives \(r_1 = 1\). Both \(r_1\) and \(r_2\) measure the dependence between \(w_1\) and \(w_2\), but from different directions.
Theorem A1
Proof
We then have
Theorem A2
Proof
Let us substitute \(r_i\)’s into formulas (7)–(10). Let us also substitute the following definitions:
knowing that:
then substitute the above into the definition of \(r_1\), we get
knowing that \(Pr[\bar{w_2}]=1-Pr[w_2]\), then:
The importance of this theorem is that our way of calculating the joint bpa \(\psi [h_1, h_2]\) is sound in that it gives a generalization of the joint probability distribution of the trustworthiness of two (potentially) dependent sources. This also follows Shafer’s general guide on how to handle non-independent evidence sources in DS [26], although Shafer did not provide the specific formulations.
B Belief Calculation Algorithm
The main algorithm is DsCorr (Algorithm 1). This function takes GraphSet which is a set of correlation graph segments. It iterates on each graph, and returns a set of the graph segments sorted by the belief of the sink node (or highest sink node for multiple sinks) in descending order.

Algorithm ComputeNodeBelief (Algorithm 2) takes a node and returns the belief value of it. There are three cases to consider for the node: (1) it is a source node; (2) it has only one parent node, (3) it has multiple parents. In the first case AssignBpaValues is called to compute the basic probability assignment based on the method in Sect. 3.2. This case applies to the alert nodes, e.g., node 1–5 in Fig. 1. In the second case the node has only one parent so the translation function is called. The third case for combination is done by first translating implicitly into the node and then combine.

Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Ou, X. (2016). A Bottom-Up Approach to Applying Graphical Models in Security Analysis. In: Kordy, B., Ekstedt, M., Kim, D. (eds) Graphical Models for Security. GraMSec 2016. Lecture Notes in Computer Science(), vol 9987. Springer, Cham. https://doi.org/10.1007/978-3-319-46263-9_1
Download citation
DOI: https://doi.org/10.1007/978-3-319-46263-9_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-46262-2
Online ISBN: 978-3-319-46263-9
eBook Packages: Computer ScienceComputer Science (R0)