Elsevier

Information Fusion

Volume 10, Issue 1, January 2009, Pages 51-69
Information Fusion

Force deployment analysis with generalized grammar

https://doi.org/10.1016/j.inffus.2008.05.008Get rights and content

Abstract

We analyse the key algorithms of data and information fusion from a linguistic point of view, and show that they fall into two paradigms: the primarily syntactic, and the primarily semantic. We propose an alternative grammatical paradigm which exploits the ability of grammar to combine syntactic inference with semantic representation. We generalize the concept of formal generative grammar to include multiple rule classes each having a topology and a base vocabulary. A generalized Chomsky hierarchy is defined. Analysing fusion algorithms in terms of grammatical representations, we find that most (including multiple hypothesis tracking) can be expressed in terms of conventional regular grammars. Situation analysis, however, is commonly attempted using first order predicate logic, which while expressive, is recursively enumerable and so scales badly.

We argue that the core issue in situation assessment is force deployment assessment, the extraction and scoring of hypotheses of the force deployment history, each of which is a multiresolution account of the activities, groupings and interactions of force components. The force deployment history represents these relationships at multiple levels of granularity and is expressed over time and space. We provide a grammatical approach for inferring such histories, and show that they can be estimated accurately and scalably. We employ a generalized context-free grammar incorporating both sequence and multiset productions. Elaborating [D. McMichael, G. Jarrad, S. Williams, M. Kennett, Grammatical methods for situation and threat analysis, in: Proceedings of The 8th International Conference on Information Fusion, Philadelphia, PA, 2005], a Generalized Functional Combinatory Categorial Grammar (GFCCG) is described that is both generalized and semantically functional (in that the semantics can be calculated directly from the syntax using a small number of rules). Force deployment modelling and parsing is demonstrated in naval and air defence scenarios. Simulation studies indicate that the method robustly handles the errors introduced by trackers under noisy cluttered conditions. The empirical time complexity of batch force deployment parsing is better than O(N1.5), where N is the number of track segments.

Force deployment assessments are required in real-time, and we have developed an incremental parser that keeps up with real-time data, and fulfils at Level 2 in the JDL fusion hierarchy the role that trackers fulfil at Level 1.

Introduction

The Joint Directors of Laboratories (JDL) model for data and information fusion [2] has evolved [3], [4] to encompass four levels, each describing a function required for the extraction of information from sensor data for the military command and control process. Level 0 involves signal processing, Level 1 concerns the extraction of objects such as combatants, vehicles and installations, Level 2 extracts the interrelationships between such objects, and Level 3 performs impact assessment. This paper is concerned with Level 2: situation analysis.

While there is no universally recognised and precise definition of situation assessment [5], there is some consensus that it at least involves the following aspects: (i) counter-surveillance measures, (ii) the environment (terrain, weather, etc.), (iii) socio-policitical background (permissible attrition rates, etc.), (iv) the disposition, deployment and location of forces, and (v) the structure and dynamics of the command systems. In this paper, we focus on real-time determination of force deployment and location to provide a coherent appreciation of the battle describing the relationships between the active force components.

The classical view of the situation assessment is that it is a batch process leading to an appreciation of the prospective battle. The approach we present is different: the situation assessment process provides a picture of both the present situation and its history, which can be updated in real time as new information is received.

The consumers of situation assessments are commanders and automated components within the command and control system. Experience at Level 1 has shown that commanders like to be given a clear picture of battle activity that they can explore and query. Irrelevant information should be excluded; and where possible, the information should be unequivocal. A force deployment analysis system should be able to answer such questions as:

  • what is the best interpretation of the situation now?

  • what are the other credible interpretations?

  • how has the situation developed over time?

  • what is this force component doing now? what has it been doing?

  • which force components are engaged on what activities? where are they?

From the point of view of an individual force component, a force deployment assessment describes its composition, its activities and its interactions with other components. The overall force deployment assessment describes the evolution of the composition, activities and interactions of all components of all forces.

The technology for constructing such analyses needs to be able to model the composition of force components, what they do, and how they interact. It is likely that the history of the objects extracted at Level 1 will have many possible interpretations. Force deployment analysis technology should be able to model such interpretations as hypotheses and to assign levels of belief to each of them, given observations and prior knowledge. It should provide efficient scalable algorithms for extracting the most likely hypotheses and for predicting the situation in the future. Implemented systems should be able to be programmed with expert knowledge, to learn from training data and to adapt to experience. It should be possible for domain experts to easily understand the information they provide. Finally, they should be cheap to configure and deploy, and be amenable to distributed and parallel implementation.

Early work on situation analysis, including the development of the situation calculus [6] and its application to military situation analysis [7] sought to generate a knowledge base that could represent a changing world. The system designer provides models in the form of rule sets, and some external process contributes facts to the knowledge base in real time. A reasoning system responds to queries emanating from human operators and other system components. A number of such systems have been built [8], [9], [10].

A constant problem with logical reasoning technology is the need to provide sufficient rules to allow reasoning and constrain search. One approach is to provide coherently created modular ontologies that represent aspects of the application domain expressed in such languages as UML and OWL [9], [11], [12], [13]. To support this process, Little and Rogova [14] have proposed a hierarchy of situation ontologies to facilitate modular construction.

The situation representation, and the queries upon it, are constrained by the limitations of the logic employed and by the scope of the ontology. First order logics are Turing-complete, but their expressive richness comes at the cost of decidability1, computational complexity2, uncertainty of computation time, and uncertainty over whether there is sufficient knowledge in the system to resolve novel real-time queries. However, their flexibility does allow rich interfaces to language and the ability to reason on knowledge itself [15]. Eliciting knowledge is expensive; but gradually better tools are becoming available. The problem remains that pure logical reasoning does not provide any mechanism for the resolution of genuine ambiguities. However, inductive methodologies, such as case-based reasoning, can be so equipped and have been applied to fusion [16].

Graphical models and Bayesian networks are able to represent uncertainty within systems of variables with dependency relationships [17]. Even though inference under such models scales badly [18], [19], [20] they can be useful when their internal variables are already bound to evidence in the outside world. Bayesian networks have been taken up within the data and information fusion community [21], [22] where mechanisms for composing networks from modules [23], [24], [25], [26] have eased their application. However, these approaches tend to require manual binding of network variables to external evidence. In situations where the objects and associations are transitory, manual binding techniques require significant configuration effort at the time of use.

A mechanism for assigning probabilities to sentences in propositional logic was first provided by Nilsson in 1986 [27]. The extension of probabilistic reasoning to first order logic has developed gradually [27], [28], [29], [30], [31], [32], [33], and recently Laskey [34], [35] has provided a comprehensive approach. While the development of these tools is a major achievement, inference using such approaches will scale even worse than the observed exponential scalability of non-probabilistic theorem proving systems [36].

Modelling the behaviour of targets has been tackled using dynamic Bayesian networks [37], which have been modularised [38], and efficient implementations have been developed [39], [40], [41], [42], [43], together with appropriate learning algorithms [44]. Applications of hidden Markov (HMM) modelling techniques are still under development, even though these cannot model long-range dependencies without large parametric sensitivities [45]. Higher-order HMMs, which attempt to capture longer range relationships tend to work poorly in comparison to context-free grammars [46] because implementations tend to have vastly too many parameters for satisfactory estimation from realistically-sized data sets. Saul and Jordan [47] have sought to alleviate this problem by modelling the high-order conditional distributions with finite mixtures of low-order components. Context-free grammars are likely to scale still better because they are composable and therefore may be able to cover a large part of the high-order state space with a relatively small number of rules.

Maupin and Jousselme [48] and Steinberg [49] have proposed belief-based logical frameworks for Level 2 analysis. Belief-based techniques for template matching under uncertainty have been developed by Yu et al. [50], [51], and a belief-based target aggregation algorithm is summarised by Bakert and Losiewicz [52].

To summarise: on one hand, unconstrained first-order predicate logic provides great flexibility — more than is required to provide force deployment assessments. As a result, tools that use it can be richly featured, but may scale badly. (Specifically, first order binary predicate logic belongs to complexity class RE, and is therefore intractable). While probabilistic logic offers better robustness and is able to provide useful marginal and conditional distributions, it scales even worse than deterministic logic. Maximisation over distributions (e.g. to find the best assessment) scales worse still [53]. There are, however, effective and scalable techniques for solving the temporal or the force aggregation sub-problems, but it appears that hitherto there has been no scalable solution for the whole.

In our analysis of the requirements of situation analysis, we suggested that a force deployment assessment is a spatio-temporal model that specifies the evolution of force structure and the activities of its components throughout the episode of interest. Such a model should describe the situation at the granularity needed at each level in the force structure. The key questions of concern to the user relate either to particular hypotheses of the overall situation or to the credibility of statements about aspects of the situation across all hypotheses. In probabilistic terms, the user needs firstly, a set of discrete situation hypotheses with a probability for each, and secondly, the ability to obtain marginal probabilities of occurrences of activities and structures at any time within the entire battle. Each overall force deployment hypothesis is a set of facts within a knowledge base and can be regarded as a possible world [54] constructed so as to make it easy to extract responses to the most important queries.

Following the earlier work of McMichael et al. [1], [55], we show that an approach based on a context-free grammar (CFG) is capable of satisfying this requirement. We give an algorithm that scalably finds the maximum a posteriori probability (MAP) force deployment assessment both in batch and incrementally. CFGs belong to the computational complexity class CFL, which in turn belongs to class P, the class of decision problems solvable in polynomial time by a Turing machine. We provide a scalable search algorithm for extracting alternative situation hypotheses in order of decreasing probability. Situation parsing is a statistical approach, the raw data for which are segments of object tracks; its output is a set of situation hypotheses each coded as a conjunction of first order predicates (with arities typically of eight or more). Marginal probabilities of actions by force structure components can be calculated by the inside–outside algorithm [56], at somewhat greater computational cost.

The Generalized Functional Combinatory Categorial Grammar (GFCCG) is designed so that the situation semantics can be derived directly from the syntactic model extracted by the parser–without recourse to semantic attachments.

Superficially, the process of rule instantiation in a first order logic reasoning system is similar to instantiating a context-free grammar rule within a parser, but there is a fundamental difference. The resulting object in a parser spans a set of tokens (the leaf nodes of the tree of which it is the root) and cannot be combined with another object that spans any of them. The effect of this constraint is that leaf nodes (the input data) are only ever used once within a parse tree. This acts as a strong constraint on the number of possible parses supported by a given set of tokens, and therefore on the computational complexity of optimal parsing.

Reasoning with context-free grammar is able to scale much better than reasoning directly with first order logic is because its objects span their leaf tokens (Section 3.1) thus restricting the number of possible hypotheses. This greatly speeds forward-chaining approaches, such as the A* algorithm we use. Backward chaining approaches are inappropriate because the goal is to extract a good force deployment model, rather than to prove a specific conjecture. Other benefits of the approach are that (i) the extracted MAP force deployment assessment is coherent over the surveillance interval and region; (ii) the estimate is robust because the entire situation is modelled and optimised simultaneously; and (iii) it avoids the construction of joint distributions of large numbers of variables.

In Section 2, we explore the analogy between linguistics and fusion. Generalized grammar, which allows operations on multiset data, is introduced in Section 3. It is then used to map the core algorithms of data and information fusion onto the levels of a generalized Chomsky hierarchy. Section 4 sets out the semantics of the proposed situation model; Section 5 describes the grammar and Section 7 presents an appropriate parser and gives some simulation results.

Section snippets

Data fusion and linguistics

The root operations of data and information fusion at Levels 1 and 2 are detection, association and state estimation. Its outcome is information. Drawing an analogy with computational linguistics, these root operations are syntactic and their product is semantics. At Level 1, the emphasis has been on complex syntax, for example, using sophisticated trackers, data association algorithms and advanced estimators of all kinds, but relatively simple semantics (such as associations between contacts

Formal grammar and the generalized Chomsky hierarchy

In the domain of discrete sequences, Chomsky defined a hierarchy of generative formal grammars with progressively greater representational power. They are respectively: regular grammar, context-free grammar, context-sensitive grammar and unconstrained grammar [58]. In generalising these concepts to the domain of spatial data and association problems we relax the requirement that the data be sequential. In linguistics, the distinction between levels in the hierarchy is important in

The semantics of military situations

An ontology is a formal specification of what objects and classes exist within a domain and the relationships that can occur between them. It comprises a set of concepts comprising predicates, functions and constants, and a set of axioms which are logical statements valid within the domain [66]. Concepts can have attributes.

While it is tempting to create large versatile ontologies, this is expensive and can give poor performance. Parsimony tends to offer greater observability and accuracy [67].

Grammatical situation analysis

In the introduction, we proposed that the crux of situation analysis is estimating the spatio-temporal structure of force deployment from track data. A natural representation is a sequence-set context-free grammar in which force-structure and temporal components are composable and the leaf nodes are track segments. The ground for estimating such structures is a probabilistic model; it facilitates scalable search for the optimum parse, and is discussed in Section 5.4.

Air situation analysis

The sequence-set GFCCG we have described was applied to an air defence scenario. The scenario consisted of up to three blue force combat air patrols (CAPs) defending a power station located near Kathrine in the Northern Territory of Australia. An instance of the scenario is shown in Fig. 8, Fig. 9. The defending fighter pairs fly from bases in Darwin on the coast and Tindal near Kathrine. A combat air patrol is a defensive formation in which two aircraft with forward-looking radar fly in a

Situation parsing

The worst case bound on computational complexity of multiset parsing using dynamic programming is O(3N) [75], where N is the number of leaf nodes. On this basis, one might therefore give up the idea of trying to parse force deployments without even trying it. However, historically, such bounds have been very conservative, to the extent of being practically useless in estimating the time complexity of a real parsers [76]. Our experiments show an empirical complexity of better than O(N1.5) for

Conclusions and discussion

We have extended the concept of grammar by generalizing its domain of application from sequences to a set of arbitrary topologies. Employing this generalization, we have demonstrated that a sequence-set grammar can be applied to scalably estimate force deployments in potentially large battles. The generalisation does not affect the rule structures, and the rule classifications of the Chomsky hierarchy still apply. A range of core inference problems and algorithms that occur within the domains

Acknowledgements

The authors are grateful to Andrew Lampert, Bella Robinson and Andrew Cunningham for the work on the GUI and interfaces to it, to Bob Lobbia, Dean Webb, Ken Manus of The Boeing Company and John Colton of CSIRO for their support and encouragement in this work.

References (86)

  • D.A. Lambert

    Assessing situations

  • C.J. Matheus, J.J. Letkowski, M.M. Kokar, C. Call, K. Baclawski, M. Hinman, J. Salerno, D. Boulware, Lessons learned...
  • C. Nowak, D. Lambert, The semantic challenge for situation assessment, in: Proceedings of The 8th International...
  • P.R. Smart, N.R. Shadbolt, L.A. Carr, M.C. Schraefel, Knowledge-based information fusion for improved situational...
  • K. Wallenius, Support for situation awareness in command and control, in: Proceedings of The 7th International...
  • M.M. Kokar, C.J. Matheus, K. Baclawski, J.A. Letkowski, M. Hinman, J. Salerno, Use cases for ontologies in information...
  • A.V. Smirnov, M.P. Pashkin, N.G. Chilov, T.V. Levashova, A.A. Krizhanovsky, Ontology-driven information integration for...
  • C. Nowak, D. Lambert, Ontology meta-model for building a situational picture of catastrophic events, in: Proceedings of...
  • J.Y. Halpern et al.

    Reasoning about Knowledge

    (1995)
  • G. Jakobson, L. Lewis, J. Buford, An approach to integrated cognitive fusion, in: Proceedings of The 7th International...
  • S. Lauritzen et al.

    Local computations with probabilities on graphical structures and their application to expert systems

    J. Roy. Stat. Soc. B

    (1988)
  • J. Paris, The Uncertain Reasoner’s Companion, Cambridge,...
  • C. Sutton, C. Morrison, P.R. Cohen, J. Moody, J. Adibi, A Bayesian blackboard for information fusion, in: Proceedings...
  • C. Howard, M. Stumptner, Using object oriented probabilistic relational models, in: Proceedings of The 8th...
  • K.B. Laskey, S.M.Mahoney, Network fragments: Representing knowledge for constructing probabilistic models, in:...
  • D. Koller et al.

    Object-oriented Bayesian networks

  • A. Pfeffer et al.

    SPOOK: A system for probabilistic object-oriented knowledge representation

  • H. Langseth et al.

    Fusion of domain knowledge with data for structural learning in object oriented domains

    J. Mach. Learn. Res.

    (2003)
  • D. Poole, Representing Bayesian networks within probabilistic Horn abduction, in: Proceedings of the Seventh Conference...
  • M. Wellman, J.S. Breese, R.P. Goldman, From knowledge bases to decision models, Knowledge Engineering Review...
  • S. Muggleton

    Stochastic logic programs

  • N. Friedman et al.

    Learning probabilistic relational models

  • B. Taskar et al.

    Discriminative probabilistic models for relational data

  • K. Laskey, First-order Bayesian logic, Technical Report, Department of Systems Engineering and Operations Research,...
  • P. Domingos et al.

    Markov logic: a unifying framework for statistical relational learning

  • I. Horrocks

    Using an expressive description logic: FaCT or fiction?

  • T. Dean et al.

    A model for reasoning about persistence and causation

    Comput. Intell.

    (1989)
  • N. Friedman, D. Koller, A. Pfeffer, Structured representation of complex stochastic systems, in: Proceedings of the...
  • K. Kanazawa et al.

    Stochastic simulation algorithms for dynamic probabilistic networks

  • X. Boyen, D. Koller, Exploiting the architecture of dynamic systems, in: Proceedings of the 16th National Conference on...
  • A. Doucet, N. de Freitas, K. Murphy, S. Russell, Rao-blackwellised particle filtering for dynamic Bayesian networks,...
  • K. Murphy, M. Paskin, Linear time inference in hierarchical HMMs, in: Proceedings of Neural Information Processing...
  • S. Sanghai et al.

    Relational dynamic Bayesian networks

    J. Artif. Int. Res.

    (2005)
  • Cited by (0)

    View full text