Force deployment analysis with generalized grammar
Introduction
The Joint Directors of Laboratories (JDL) model for data and information fusion [2] has evolved [3], [4] to encompass four levels, each describing a function required for the extraction of information from sensor data for the military command and control process. Level 0 involves signal processing, Level 1 concerns the extraction of objects such as combatants, vehicles and installations, Level 2 extracts the interrelationships between such objects, and Level 3 performs impact assessment. This paper is concerned with Level 2: situation analysis.
While there is no universally recognised and precise definition of situation assessment [5], there is some consensus that it at least involves the following aspects: (i) counter-surveillance measures, (ii) the environment (terrain, weather, etc.), (iii) socio-policitical background (permissible attrition rates, etc.), (iv) the disposition, deployment and location of forces, and (v) the structure and dynamics of the command systems. In this paper, we focus on real-time determination of force deployment and location to provide a coherent appreciation of the battle describing the relationships between the active force components.
The classical view of the situation assessment is that it is a batch process leading to an appreciation of the prospective battle. The approach we present is different: the situation assessment process provides a picture of both the present situation and its history, which can be updated in real time as new information is received.
The consumers of situation assessments are commanders and automated components within the command and control system. Experience at Level 1 has shown that commanders like to be given a clear picture of battle activity that they can explore and query. Irrelevant information should be excluded; and where possible, the information should be unequivocal. A force deployment analysis system should be able to answer such questions as:
- •
what is the best interpretation of the situation now?
- •
what are the other credible interpretations?
- •
how has the situation developed over time?
- •
what is this force component doing now? what has it been doing?
- •
which force components are engaged on what activities? where are they?
From the point of view of an individual force component, a force deployment assessment describes its composition, its activities and its interactions with other components. The overall force deployment assessment describes the evolution of the composition, activities and interactions of all components of all forces.
The technology for constructing such analyses needs to be able to model the composition of force components, what they do, and how they interact. It is likely that the history of the objects extracted at Level 1 will have many possible interpretations. Force deployment analysis technology should be able to model such interpretations as hypotheses and to assign levels of belief to each of them, given observations and prior knowledge. It should provide efficient scalable algorithms for extracting the most likely hypotheses and for predicting the situation in the future. Implemented systems should be able to be programmed with expert knowledge, to learn from training data and to adapt to experience. It should be possible for domain experts to easily understand the information they provide. Finally, they should be cheap to configure and deploy, and be amenable to distributed and parallel implementation.
Early work on situation analysis, including the development of the situation calculus [6] and its application to military situation analysis [7] sought to generate a knowledge base that could represent a changing world. The system designer provides models in the form of rule sets, and some external process contributes facts to the knowledge base in real time. A reasoning system responds to queries emanating from human operators and other system components. A number of such systems have been built [8], [9], [10].
A constant problem with logical reasoning technology is the need to provide sufficient rules to allow reasoning and constrain search. One approach is to provide coherently created modular ontologies that represent aspects of the application domain expressed in such languages as UML and OWL [9], [11], [12], [13]. To support this process, Little and Rogova [14] have proposed a hierarchy of situation ontologies to facilitate modular construction.
The situation representation, and the queries upon it, are constrained by the limitations of the logic employed and by the scope of the ontology. First order logics are Turing-complete, but their expressive richness comes at the cost of decidability1, computational complexity2, uncertainty of computation time, and uncertainty over whether there is sufficient knowledge in the system to resolve novel real-time queries. However, their flexibility does allow rich interfaces to language and the ability to reason on knowledge itself [15]. Eliciting knowledge is expensive; but gradually better tools are becoming available. The problem remains that pure logical reasoning does not provide any mechanism for the resolution of genuine ambiguities. However, inductive methodologies, such as case-based reasoning, can be so equipped and have been applied to fusion [16].
Graphical models and Bayesian networks are able to represent uncertainty within systems of variables with dependency relationships [17]. Even though inference under such models scales badly [18], [19], [20] they can be useful when their internal variables are already bound to evidence in the outside world. Bayesian networks have been taken up within the data and information fusion community [21], [22] where mechanisms for composing networks from modules [23], [24], [25], [26] have eased their application. However, these approaches tend to require manual binding of network variables to external evidence. In situations where the objects and associations are transitory, manual binding techniques require significant configuration effort at the time of use.
A mechanism for assigning probabilities to sentences in propositional logic was first provided by Nilsson in 1986 [27]. The extension of probabilistic reasoning to first order logic has developed gradually [27], [28], [29], [30], [31], [32], [33], and recently Laskey [34], [35] has provided a comprehensive approach. While the development of these tools is a major achievement, inference using such approaches will scale even worse than the observed exponential scalability of non-probabilistic theorem proving systems [36].
Modelling the behaviour of targets has been tackled using dynamic Bayesian networks [37], which have been modularised [38], and efficient implementations have been developed [39], [40], [41], [42], [43], together with appropriate learning algorithms [44]. Applications of hidden Markov (HMM) modelling techniques are still under development, even though these cannot model long-range dependencies without large parametric sensitivities [45]. Higher-order HMMs, which attempt to capture longer range relationships tend to work poorly in comparison to context-free grammars [46] because implementations tend to have vastly too many parameters for satisfactory estimation from realistically-sized data sets. Saul and Jordan [47] have sought to alleviate this problem by modelling the high-order conditional distributions with finite mixtures of low-order components. Context-free grammars are likely to scale still better because they are composable and therefore may be able to cover a large part of the high-order state space with a relatively small number of rules.
Maupin and Jousselme [48] and Steinberg [49] have proposed belief-based logical frameworks for Level 2 analysis. Belief-based techniques for template matching under uncertainty have been developed by Yu et al. [50], [51], and a belief-based target aggregation algorithm is summarised by Bakert and Losiewicz [52].
To summarise: on one hand, unconstrained first-order predicate logic provides great flexibility — more than is required to provide force deployment assessments. As a result, tools that use it can be richly featured, but may scale badly. (Specifically, first order binary predicate logic belongs to complexity class RE, and is therefore intractable). While probabilistic logic offers better robustness and is able to provide useful marginal and conditional distributions, it scales even worse than deterministic logic. Maximisation over distributions (e.g. to find the best assessment) scales worse still [53]. There are, however, effective and scalable techniques for solving the temporal or the force aggregation sub-problems, but it appears that hitherto there has been no scalable solution for the whole.
In our analysis of the requirements of situation analysis, we suggested that a force deployment assessment is a spatio-temporal model that specifies the evolution of force structure and the activities of its components throughout the episode of interest. Such a model should describe the situation at the granularity needed at each level in the force structure. The key questions of concern to the user relate either to particular hypotheses of the overall situation or to the credibility of statements about aspects of the situation across all hypotheses. In probabilistic terms, the user needs firstly, a set of discrete situation hypotheses with a probability for each, and secondly, the ability to obtain marginal probabilities of occurrences of activities and structures at any time within the entire battle. Each overall force deployment hypothesis is a set of facts within a knowledge base and can be regarded as a possible world [54] constructed so as to make it easy to extract responses to the most important queries.
Following the earlier work of McMichael et al. [1], [55], we show that an approach based on a context-free grammar (CFG) is capable of satisfying this requirement. We give an algorithm that scalably finds the maximum a posteriori probability (MAP) force deployment assessment both in batch and incrementally. CFGs belong to the computational complexity class CFL, which in turn belongs to class P, the class of decision problems solvable in polynomial time by a Turing machine. We provide a scalable search algorithm for extracting alternative situation hypotheses in order of decreasing probability. Situation parsing is a statistical approach, the raw data for which are segments of object tracks; its output is a set of situation hypotheses each coded as a conjunction of first order predicates (with arities typically of eight or more). Marginal probabilities of actions by force structure components can be calculated by the inside–outside algorithm [56], at somewhat greater computational cost.
The Generalized Functional Combinatory Categorial Grammar (GFCCG) is designed so that the situation semantics can be derived directly from the syntactic model extracted by the parser–without recourse to semantic attachments.
Superficially, the process of rule instantiation in a first order logic reasoning system is similar to instantiating a context-free grammar rule within a parser, but there is a fundamental difference. The resulting object in a parser spans a set of tokens (the leaf nodes of the tree of which it is the root) and cannot be combined with another object that spans any of them. The effect of this constraint is that leaf nodes (the input data) are only ever used once within a parse tree. This acts as a strong constraint on the number of possible parses supported by a given set of tokens, and therefore on the computational complexity of optimal parsing.
Reasoning with context-free grammar is able to scale much better than reasoning directly with first order logic is because its objects span their leaf tokens (Section 3.1) thus restricting the number of possible hypotheses. This greatly speeds forward-chaining approaches, such as the A* algorithm we use. Backward chaining approaches are inappropriate because the goal is to extract a good force deployment model, rather than to prove a specific conjecture. Other benefits of the approach are that (i) the extracted MAP force deployment assessment is coherent over the surveillance interval and region; (ii) the estimate is robust because the entire situation is modelled and optimised simultaneously; and (iii) it avoids the construction of joint distributions of large numbers of variables.
In Section 2, we explore the analogy between linguistics and fusion. Generalized grammar, which allows operations on multiset data, is introduced in Section 3. It is then used to map the core algorithms of data and information fusion onto the levels of a generalized Chomsky hierarchy. Section 4 sets out the semantics of the proposed situation model; Section 5 describes the grammar and Section 7 presents an appropriate parser and gives some simulation results.
Section snippets
Data fusion and linguistics
The root operations of data and information fusion at Levels 1 and 2 are detection, association and state estimation. Its outcome is information. Drawing an analogy with computational linguistics, these root operations are syntactic and their product is semantics. At Level 1, the emphasis has been on complex syntax, for example, using sophisticated trackers, data association algorithms and advanced estimators of all kinds, but relatively simple semantics (such as associations between contacts
Formal grammar and the generalized Chomsky hierarchy
In the domain of discrete sequences, Chomsky defined a hierarchy of generative formal grammars with progressively greater representational power. They are respectively: regular grammar, context-free grammar, context-sensitive grammar and unconstrained grammar [58]. In generalising these concepts to the domain of spatial data and association problems we relax the requirement that the data be sequential. In linguistics, the distinction between levels in the hierarchy is important in
The semantics of military situations
An ontology is a formal specification of what objects and classes exist within a domain and the relationships that can occur between them. It comprises a set of concepts comprising predicates, functions and constants, and a set of axioms which are logical statements valid within the domain [66]. Concepts can have attributes.
While it is tempting to create large versatile ontologies, this is expensive and can give poor performance. Parsimony tends to offer greater observability and accuracy [67].
Grammatical situation analysis
In the introduction, we proposed that the crux of situation analysis is estimating the spatio-temporal structure of force deployment from track data. A natural representation is a sequence-set context-free grammar in which force-structure and temporal components are composable and the leaf nodes are track segments. The ground for estimating such structures is a probabilistic model; it facilitates scalable search for the optimum parse, and is discussed in Section 5.4.
Air situation analysis
The sequence-set GFCCG we have described was applied to an air defence scenario. The scenario consisted of up to three blue force combat air patrols (CAPs) defending a power station located near Kathrine in the Northern Territory of Australia. An instance of the scenario is shown in Fig. 8, Fig. 9. The defending fighter pairs fly from bases in Darwin on the coast and Tindal near Kathrine. A combat air patrol is a defensive formation in which two aircraft with forward-looking radar fly in a
Situation parsing
The worst case bound on computational complexity of multiset parsing using dynamic programming is [75], where N is the number of leaf nodes. On this basis, one might therefore give up the idea of trying to parse force deployments without even trying it. However, historically, such bounds have been very conservative, to the extent of being practically useless in estimating the time complexity of a real parsers [76]. Our experiments show an empirical complexity of better than for
Conclusions and discussion
We have extended the concept of grammar by generalizing its domain of application from sequences to a set of arbitrary topologies. Employing this generalization, we have demonstrated that a sequence-set grammar can be applied to scalably estimate force deployments in potentially large battles. The generalisation does not affect the rule structures, and the rule classifications of the Chomsky hierarchy still apply. A range of core inference problems and algorithms that occur within the domains
Acknowledgements
The authors are grateful to Andrew Lampert, Bella Robinson and Andrew Cunningham for the work on the GUI and interfaces to it, to Bob Lobbia, Dean Webb, Ken Manus of The Boeing Company and John Colton of CSIRO for their support and encouragement in this work.
References (86)
Computational complexity of probabilistic inference using Bayesian belief networks (research note)
Artif. Int.
(1990)On the hardness of approximate reasoning
Artif. Int.
(1996)Probabilistic logic
Artif. Int.
(1986)Probabilistic Horn abduction and Bayesian networks
Artif. Int.
(1993)- D. McMichael, G. Jarrad, S. Williams, M. Kennett, Grammatical methods for situation and threat analysis, in:...
- F. White, A model for data fusion, in: Proceedings of the 1st National Symposium on Sensor Fusion,...
- A. Steinberg, C. Bowman, F. White, Revisions to the JDL data fusion model, in: Proceedings of the AeroSense, vol. 3719,...
- J. Llinas, C. Bowman, G. Rogova, A. Steinberg, E. Waltz, F. White, Revisiting the JDL data fusion model II, in:...
- et al.
Multisensor Data Fusion
(1990) - et al.
Situations and Attitudes
(1983)