Vous êtes sur la page 1sur 10

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/317239986

A Process for Human-aided Multi-Entity Bayesian Networks Learning in Predictive


Situation Awareness

Conference Paper · May 2016

CITATIONS READS

4 41

4 authors:

Cheol Young Park Kathryn Blackmond Laskey


George Mason University George Mason University
19 PUBLICATIONS   69 CITATIONS    238 PUBLICATIONS   5,950 CITATIONS   

SEE PROFILE SEE PROFILE

Paulo Costa Shou Matsumoto


George Mason University George Mason University
136 PUBLICATIONS   1,325 CITATIONS    17 PUBLICATIONS   143 CITATIONS   

SEE PROFILE SEE PROFILE

Some of the authors of this publication are also working on these related projects:

ADS-B Security View project

ITA/GMU Distributed Simulation Project View project

All content following this page was uploaded by Kathryn Blackmond Laskey on 07 September 2017.

The user has requested enhancement of the downloaded file.


A Process for Human-aided Multi-Entity Bayesian
Networks Learning in Predictive Situation Awareness
Cheol Young Park, Kathryn Blackmond Laskey, Paulo C. G. Costa, Shou Matsumoto
The Sensor Fusion Lab & Center of Excellence in C4I
George Mason University, MS 4B5
Fairfax, VA 22030-4444 U.S.A.
cparkf@masonlive.gmu.edu, [klaskey, pcosta]@gmu.edu, smatsum2@masonlive.gmu.edu

Abstract— Predictive Situation Awareness (PSAW) is the and activity of groups of targets, and generated alerts [3].
ability to estimate and predict aspects of a temporally evolving PROGNOS represented domain knowledge using probabilistic
situation. PSAW systems reason about complex and uncertain ontologies, fused the knowledge dynamically into complex
situations involving multiple targets observed by multiple sensors situation models, and inferred aspects of the current situation
at different times. Multi-Entity Bayesian Networks (MEBN) are from the models [6]. Laskey [9] described a MEBN application
rich enough to represent and reason about uncertainty in for using a model of user behavior to detect insider threats
complex, knowledge-rich domains, and have been applied to from document retrieval patterns. Suzic [5] introduced a
representation and reasoning for PSAW. To overcome a labor- generic framework for plan recognition using MEBN. The
intensive and insufficiently agile process for manual MEBN
framework focused on estimating plans of multi-agent
modeling by a domain expert, MEBN machine learning was
organizations through time using an ontology that supports
developed. Although technologies for machine learning have
improved dramatically, the necessary capabilities to build a MEBN and contained four main template random variables
MEBN model efficiently do not yet exist. The search space for (i.e., utility, plan, state, and observation).
components of an MEBN model is too large and complex to For efficiently designing a MEBN model, a reference
investigate all possible structures, variables, and parameters. For architecture [32] and a structured process [25] for constructing
this reason, this paper proposes a method which relies partially semantically rich models for reasoning under uncertainty in
on expert knowledge and insight to reduce the search space. The
complex environments were suggested. Recently, a reference
proposed method, a process for Human-aided MEBN learning in
MEBN model for PSAW, which offers referential elements of
PSAW, is a framework to develop a MEBN model from the
domain expert's knowledge combined with relational data. This
MEBN for PSAW, was proposed [15].
paper presents the process for Human-aided MEBN learning in Manual MEBN modeling by a domain expert is a labor-
PSAW and a case study to evaluate the process on development intensive and insufficiently agile process. Therefore, greater
of a defense system in PSAW. automation through machine learning method may enhance
agility. For this reason, a basic MEBN learning method was
Keywords—Data Fusion; Situation Awareness; Predictive
suggested [18][19]. Although technologies for machine
Situation Awareness; Bayesian Networks; Multi-Entity Bayesian
learning have improved dramatically, the methods to learn a
Networks; Modeling Process; Human-aided Machine Learning
MEBN model from data have not yet been reported. The
I. INTRODUCTION search space for building the model is too large and complex to
investigate all possible structures, variables, and parameters.
Predictive Situation Awareness (PSAW) requires For this reason, the proposed method relies partially on expert
combining information from multiple sensors to reason about knowledge and insight to reduce the search space. A process
multiple targets over a range of times. Furthermore, the number for Human-aided MEBN learning in PSAW is a framework to
of entities and the relationships between them may be uncertain. develop an MEBN model from the domain expert's knowledge
For this reason, PSAW requires an expressive language for as well as data.
representing and reasoning about situations. Multi-Entity
Bayesian Networks (MEBN) [1] is such an expressive This paper (1) provides background information for the
language and has been applied to PSAW systems proposed process, (2) introduces the process for Human-aided
[2][3][4][5][6][7]. MEBN combines First-Order Logic with MEBN learning in PSAW, and (3) presents a case study on
Bayesian Networks to represent and reason under uncertainty developing a defense system in PSAW.
in complex, knowledge-rich domains. The increased expressive
and reasoning power over ordinary Bayesian Networks (BNs) II. BACKGROUND
[8] is an advantage for PSAW. In this section, a definition of PSAW is presented and
MEBN is introduced as a representation formalism for PSAW.
MEBN has been applied to several PSAW systems.
TSGUDA fused data from different sensors, identified the type Human-aided MEBN learning uses a relational model (RM) as
a data schema for a dataset. We choose RM because it is the For example, two vehicles V1 and V2 may perform certain
most popular database model and has the necessary expressive activities at time T1. The meaning of the situation (MS) means
power to represent entities and their relationships. It is a set of the activities stimulated by the actors. For example, if
necessary to define how to convert elements of RM to elements the two vehicles’ activities are attacking, the meaning of the
of MEBN, so a mapping rule between MEBN and RM, called situation (MS) may be an attack situation.
MEBN-RM, is introduced. Also, we introduce a reference
MEBN model for PSAW which enables us to easily develop a B. Multi-Entity Bayesian Networks
MEBN model for PSAW by supporting the design of a MEBN MEBN allows compact representation of repeated structure
model for PSAW. The process for Human-aided MEBN in a joint distribution on a set of random variables. MEBN
learning in PSAW, which this paper proposes, is a modification represents domain knowledge using an MTheory, a collection
of a probabilistic ontology modeling process, called of MFrags. An MFrag is a fragment of a graphical model that
Uncertainty Modeling Process for Semantic Technology is a template for probabilistic relationships among instances of
(UMP-ST), so UMP-ST is introduced in this section. its random variables (RVs). RVs in an MFrag can contain
ordinary variables, which can be instantiated for different
A. Predictive Situation Awareness domain entities to create multiple instances of the RVs. We can
According to the most commonly cited definition, situation think of an MFrag as a class that can generate instances of BN
awareness (SAW) is composed of three processes: perception fragments. These can be assembled into a Bayesian network,
of elements in the environment, comprehension of the current called a situation-specific Bayesian Network (SSBN). A given
situation, and projection of the future status [10][11]. Breton MTheory can represent many different SSBNs, depending on
and Rousseau [12] classified 26 SAW definitions and the specific entity instances present in the situation. The
identified a set of common elements of SAW. They identified following definition is taken from [1].
two distinct varieties, which they termed State- and Process-
oriented SAW. In their definition, Process-oriented SAW Definition (MFrag). An MFrag F consists of: (i) a set 𝑪 of
focuses on the link between the situation and the cognitive context nodes, which represent conditions under which the
processes generating SAW, while State-oriented SAW focuses distribution defined in the MFrag is valid; (ii) a set 𝑰 of input
on the link between the situation and an internal representation nodes, which have their distributions defined elsewhere and
of elements present in the situation. condition the distributions defined in the MFrag; (iii) a set 𝑹 of
resident nodes, whose distributions are defined in the MFrag1;
Predictive Situation Awareness (PSAW) emphasizes the (iv) an acyclic directed graph G, whose nodes are associated
ability to make predictions about aspects of a temporally with resident and input nodes; and (iv) a set 𝑳 of local
evolving situation. Traditionally, humans have been distributions, in which an element of 𝑳 is associated with each
responsible for higher-level fusion of low-level fusion outputs resident node. Each input node must be a root node of G and
to estimate and predict the evolving situation. A PSAW system may have only resident nodes as children. The RVs in an
must aggregate state estimates provided by lower level MFrag may have ordinary variables as arguments. We can
information fusion (LLIF) systems to help users understand substitute different domain entities for the ordinary variables to
key aspects of the situation and project its likely evolution. In make instances of the RVs in the MFrag.
order to more effectively support humans in understanding and
predicting a complex situation, a PSAW system needs a An MTheory M is a collection of MFrags that satisfies
semantically rich representation to capture attributes of, conditions given in [1] ensuring the existence of a unique joint
relationships among, and processes associated with various distribution over its random variables. Like other expressive
kinds of entities in a situation. It also needs to reason under probabilistic languages such as Object-oriented Bayesian
uncertainty about the situation. We use the following, taken networks [13] or probabilistic relational models [14], MEBN
from [15] as our definition of PSAW. allows random variables to be defined as templates than can be
repeatedly instantiated to construct probabilistic models with
Definition (Predictive Situation Awareness). Predictive repeated structure.
Situation Awareness (PSAW) is the ability to estimate and
predict a possible situation involving multiple actors (Atr) C. Relational Model (RM)
and/or objects (Obj) in different locations (Loc), in which In 1969, Codd proposed the Relational Model (RM) as a
actors may trigger events (Evt) or activities (Act) occurring database model based on first-order predicate logic [23][24].
over time (T), and where the meaning of the situation (MS) is RM is the most popular database model. A relational database
revealed by integrating previous knowledge with evidence (RDB) is a database that uses RM to describe and organize
from multiple sources. data. In RM, data are organized as a collection of relations. A
relation is an abstract definition of a class of entities or a
Actors (Atr) are entities which can generate events (Evt) or relationship that can hold between classes of entities. An
activities (Act), whereas objects (Obj) can be involved in instance of a relation is depicted as a table in which each
events or activities but cannot actively initiate them. For column is an attribute of the relation and each row, also called
example, vehicles V1 and V2 can be modeled as actors, and a tuple, contains the value of each attribute for an individual
they may be able to perform Move, Defend, or Attack activities. entity of the class represented by the relation. An entry in the
(In a richer representation, we might model the vehicles as table, called a cell, is the value of the attribute associated with
objects and their drivers as actors.) On the other hand, a region the column for the entity associated with the row. A key for a
or stone is an object and cannot perform an activity. Events and
activities happen at a given time or during a time interval (T). 1
Bold italic letters are used to denote sets.
relation is one or more attributes that uniquely identify a argument of the function VehicleType is the primary key of the
particular domain entity or row. A primary key uniquely Vehicle relation, and the output is the cell (either Tracked or
identifies the individual entities in the relation. A foreign key Wheeled) of the VehicleType attribute.
points to the primary key in another relation. The cardinality of
a relation is the number of rows in the table, i.e., the number of In MEBN, context nodes are used to specify constraints
under which the local distributions apply. In RM, the context
unique entities of the type represented by the relation. The
degree of the relation is the number of columns in the table, nodes are used for limiting a relation instance in a relation. For
MEBN-RM, we identify four types of the context node (Isa,
i.e., the number of attributes of entities of the type represented
by the relation. Value-Constraint, Slot-filler, and Entity-Constraint type) [19].
E. A Reference MEBN Model for PSAW
A reference MEBN model for PSAW specifies reference
MFrags, RVs, and entities which support the design of a
MEBN model for PSAW. Such a MEBN model is called a
PSAW-MTheory. The PSAW-MTheory is designed to reason
about PSAW questions and can be supported by the reference
model, called a reference PSAW-MEBM model.
To perform PSAW, we may want to answer various
questions, e.g., "Will the weather be cloudy?", "How many
Fig. 1. Example of Relations military vehicles are we going to encounter?", "How high will
the level of danger for an enterprise be?", "Where will the
Fig. 1 shows an example of a relational model. There are three target be located?", "Where will the event occur?", "What is
relations: Vehicle, Report, and Communication. Columns of the type of the target?", "What type of activity will the target
each relation represent attributes of the relation. For example, perform?", and "What will the enemy do and why?"
the Vehicle relation has two attributes: Key, which uniquely [20][4][21]. Also, Roy [22] proposed a broad spectrum of
identifies each individual vehicle, and VehicleType, which questions to be answered in situation analysis (e.g., situation
indicates whether the vehicle is tracked or wheeled. element contextual analysis).
D. MEBN-RM Mapping Model An entity in the reference model can be classified into five
MEBN-RM is a framework that maps elements of RM to categories; (1) the time entity T, (2) the observer entity OR, (3)
elements of MEBN [19]. MEBN-RM contains an entity the sensor entity SR, (4) the target entity TR, and (5) the
mapping between a relation in RM and an entity in MEBN, a interpreted target entity IT. These categories can be used to
resident node mapping between an attribute in RM and a describe a situation in which an observer observes a target and
resident node in MEBN, and a context node mapping between interprets it as an interpreted target using a sensor at a certain
an attribute in RM and a context node in MEBN. time. For example, an omnidirectional radar (as SR) controlled
by a commander (as OR) detects a vehicle object (as TR) and
In MEBN, an entity is a unique thing with independent produces a vehicle object report (as IT) at time T1 (as T).
existence. We can distinguish between entity types and entity
instances. For example, individual people (e.g., John, Mathew) In the reference model, random variables containing
are instances of the person entity type. In MEBN-RM, an ordinary variables of type T, OR, SR, TR, or IT are called
entity relation from RM is mapped to an entity type in MEBN. PSAW-RVs. For example, a situation in which two vehicles
For example, the Vehicle relation (Fig. 1) can be mapped to a communicate with each other and are detected by a COMINT
VEHICLE entity type in MEBN. (communications intelligence) sensor at time T1 can be
described by a PSAW-RV, specified by Communicated(v1, v2,
A resident node in MEBN can be described as a Function, a cmt, t), where v1, v2, cmt, and t are ordinary variables that can
Predicate, or a Formula of First-order logic (FOL). MEBN be filled in by a first vehicle, a second vehicle, a COMINT
allows a modeler to specify a probability distribution for the sensor, and a time, respectively.
truth-value of a predicate or the value of a function. In FOL, a
predicate represents a true/false statement about entities in the In MEBN, an MFrag is a fragment of a probabilistic graph-
domain. For example, Communication(x, y) is a predicate that ical model in which the nodes contain variables that are
expresses whether the entities indicated by the arguments x and placeholders for domain entities. MFrags can be instantiated
y are communicating. In RM, we can express a predicate as a with domain entities and combined into complex models with
relation in which the attributes are arguments of the predicate repeated structure. MFrags for PSAW, called PSAW-MFrags,
(e.g., the Communication relation in Fig. 1). In FOL, a function contains PSAW-RV. The reference MEBN model for PSAW
is a mapping from domain entities to a value. For example, the suggests four groups: (1) Relation of Observed Target, (2)
function VehicleType(x) is a function that maps its argument x Sensing Conditions, (3) Relation of Interpreted Target, and (4)
to wheeled if it is a wheeled vehicle and tracked if it is a Situation. These four groups can be constructed as four MFrags:
tracked vehicle. In RM, we can express a function as an (1) a Relation of Observed Target MFrag, (2) a Sensing Condi-
attribute of a relation. A primary key in the relation is used for tions MFrag, (3) a Relation of Interpreted Target MFrag, and
argument (s) of the function and a cell of the attribute is used (4) a Situation MFrag, respectively. The Relation of Observed
for an output value of the function, so the function maps from Target MFrag defines distributions for PSAW-RVs related to
the primary key to the cell. For example, in Fig. 1, the relations of an observed target. For example, it can include the
PSAW-RVs mission, size, type, and activity of a target. The training dataset. This paper improves the previous proces by
Relation of Sensing Conditions MFrag represents probabilistic using expert knowledge to limit the search space over
knowledge about conditions of sensors. For example, it can parameters, variables, and structures. The process is a
include the PSAW-RVs performance, range, and weather modification of UMP-ST to be applied to Human-aided MEBN
condition of a sensor. The Relation of Interpreted Target learning in PSAW, called HMLP. As UMP-ST containing the
MFrag represents probabilistic knowledge about relations for four main disciplines, the presenting process contains the four
interpreted targets. For example, it can include the PSAW-RVs steps (Fig. 2): (1) Analyze Requirements, (2) Design World
reported mission, reported size, reported type, and reported Model and Rules, (3) Construct Reasoning Model, and (4) Test
activity of a target. The Situation MFrag defines distributions Reasoning Model.
for situations of targets. For example, it can include the
PSAW-RVs danger level situation, collaborating situation,
and communicating situation.
F. Uncertainty Modeling Process for Semantic Technology
UMP-ST is a process for constructing a probabilistic
ontology. A probabilistic ontology represents domain
knowledge, including uncertainty about entity attributes and Fig. 2. Process for Human-Aided MEBN Learning
relationships [4]. PR-OWL [25] extends the OWL language to
represent probabilistic ontologies. UMP-ST supports the Inputs of the process can be needs and/or missions from
construction of PR-OWL probabilistic ontologies. The UMP- stakeholders. Outputs of the process are a reasoning model (in
ST process consists of four main disciplines: (1) Requirements, our case, a PSAW MTheory). The following sub-sections
(2) Analysis & Design, (3) Implementation, and (4) Test. describe these four steps.
The Requirements discipline defines goals, queries, and A. Analyze Requirements
evidence for a probabilistic ontology. The goals are objectives
This step is to identify requirements for development of a
to be achieved by reasoning with the probabilistic ontology
reasoning model. As with requirements in UMP-ST,
(e.g., identify a ground target). The queries are specific
requirements in the HMLP define goals to be achieved, queries
questions for which the answers help to achieve the objectives.
to answer, and evidence to be used in answering queries. Also,
For example, what is the type of the target? The evidence is
the requirements should include performance criteria for
information used to answer the query (e.g., history of the speed
verification of the reasoning model. These performance criteria
of the target). The Analysis & Design discipline designs
are used in the Test Reasoning Model step. Before the Analyze
entities, attributes for the entities, relationships between the
Requirements step begins, stakeholders provide their initial
entities, and rules for attributes and relationships to represent
requirements containing needs, wants, missions, and objectives.
uncertainty. These are associated with the goals, queries, and
These initial requirements may not be defined formally.
evidence in the Requirement discipline. For example, suppose
Therefore, to clarify the initial requirements, operational
that a vehicle entity has two attributes, type, and speed. Then
scenarios are developed. In other words, the operational
an example of a rule might be that if the speed is low, the type
scenarios enable to identify the goals, queries, and evidence in
is likely to be a tracked vehicle. The Implementation discipline
the requirements.
is a step to develop a probabilistic ontology from outputs
developed in the Analysis & Design discipline. Entities, This step contains two sub-steps (Fig. 3): (1) an Identify
attributes, relationships, and rules are mapped to elements of Goals step and (1) an Identify Queries/Evidences step.
the probabilistic ontology. For example, the attributes type and
speed are mapped to random variables type and speed,
respectively. The rule for the speed and type is converted to the
joint probability for the random variables type and speed. In
the Test discipline, the probabilistic ontology developed in the
previous step is evaluated to assess its correctness. The
correctness can be measured by three approaches: (1)
Elicitation Review, in which completeness of the probabilistic
ontology addressing requirements are reviewed, (2) Importance Fig. 3. Analyze Requirements
Analysis, which is a form of sensitivity analysis that examines
the strength of influence of each random variable on other The goals represent missions of the reasoning model we are
random variables, and (3) Case-based Evaluation, in which developing. The queries are specific questions for which the
various scenarios are defined and used to examine the reasoning model is used to estimate and/or predict answers.
reasoning implications of the probabilistic ontology [16]. The evidence consists of inputs used for reasoning. From these
sub-steps, a set of goals, a set of queries for each goal, and a set
III. PROCESS FOR HUMAN-AIDED MEBN LEARNING IN of evidence for each query are defined. The following shows
PSAW an illustrative example of requirement.
A basic MEBN learning method for relational databses was Goal 1: Identify characteristics of a target.
suggested [18][19]. MEBN learning finds parameters, variables,
and structure for an MTheory that provides a good fit to a
Query 1.1: What is the speed of the target at a given contain a sensor entity SR for the position attribute, so we add
time? both a sensor entity of ground moving target indicator (GMTI)
and its performance attribute into the world model.
Evidence 1.1.1: Position report for the target.
In this step, we determine whether data can be obtained for
The performance criteria in the requirements can be the attribute, and if so, either collect data or identify an existing
specified in terms of some measure of accuracy (e.g., the Brier
dataset. We usually divide the data into a training dataset and a
score [17], the Kullback-Leibler (KL) divergence [26], and the test dataset. The training dataset is used for machine learning in
area under the receiver operating characteristic curve (AUC)).
the Construct Reasoning Model step. The test dataset is used
For example, we might require that the sum of KL divergence for verification in the Test Reasoning Model step. If no data
values between ground truth and estimated results from a
can be obtained, we use the judgment of domain experts to
reasoning model shall be less than a given threshold. The specify the necessary probability distributions. For example, a
performance criteria are determined by stakeholder agreement.
belief for the target type attribute can be P(Wheeled) = 0.8 and
B. Design World Model and Rules P(Tracked) = 0.2. If neither data nor expert judgment is
The Design World Model and Rules step develops a world available, we consider whether the attribute is really necessary.
model and rules. The world model and the rules provide a clear For this, we can return the Analyze Requirements step to
idea by which the reasoning model can be formed. This step modify the requirements.
decomposes into two sub-steps (Fig. 4): (1) a Design World The Design Rules step defines rules for the identified
Model step and (2) a Design Rules step. The Design World attributes (i.e., type, speed, previous speed, GMTI performance,
Model step defines the world model from the requirements, and position) by expert knowledge. For example, we noted
domain knowledge and/or existing data schemas. The world earlier that the type attribute most likely influences the speed
model is used to identify rules. The Design Rules step defines a attribute and the previous speed attribute also likely influences
rule or an influencing relationship between attributes (e.g., A the current speed attribute. These beliefs from expert
and B) of relations in the world model. The rules are used to knowledge become a causal relationship rule as shown the
identify a reasoning model. The influencing relationship is a following.
relationship between attributes in which there is an unknown
causality between the attributes (e.g., influencing(A, B)). If we Rule 1: causal(type, speed)
know the causality, the influencing relationship becomes a Rule 2: causal(previous speed, speed)
causal relationship (e.g., causal(A, B)).
In this step, a (conditional) local distribution for an attribute
(e.g., the speed attribute) can be defined by expert knowledge.
In some cases, there is no dataset for a rule and all we have is
expert knowledge. For example, a conditional local distribution
for the speed attribute given the type attribute can be identified
by a domain expert (e.g., if a vehicle type is wheeled, then the
speed of the vehicle on a road is normally distributed with a
mean of 50MPH and a standard deviation of 20MPH).
C. Construct Reasoning Model
Fig. 4. Design World Model and Rules The Construct Reasoning Model step develops a reasoning
model from a training dataset a world model, and rules.
The Design World Model step uses requirements, domain
knowledge and/or existing data schemas to develop a world
model. The world can be represented in a modeling language
(e.g., Entity–Relationship (ER) model, Enhanced Entity–
Relationship (EER), RM, or Unified Modeling Language
(UML)). A requirement specifies a query and evidence for the
query. The elements of the requirement are used to define
corresponding elements in the world model. For example,
suppose that the requirements specify speed (Query 1.1) and
position attributes (Evidence 1.1.1) for a target g at a time t.
Based on these requirements, these two attributes are included Fig. 5. Construct Reasoning Model
in the world model. We can then identify additional attributes
related to these attributes. For example, a type attribute for the This step decomposes into two sub-steps (Fig. 5): (1) a Map
target g most likely influences the speed attribute. The speed to Reasoning Model step and (2) a Learn Reasoning Model
attribute at the previous time probably influences the current step. The Map to Reasoning converts the world model and
speed attribute. These new attributes and relationships are also rules to a candidate reasoning model. The Learn Reasoning
included in the world model. In this step, the reference MEBN Model uses a machine learning method to learn the model from
model for PSAW can be used to identify more attributes for the a training dataset.
world model. For example, from the five entities (i.e., T, OR,
In the Map to Reasoning Model step, MEBN-RM is used
SR, TR, and IT), we can realize that the world model should
for a reference of mapping rule (i.e., how to convert RM to
MEBN). The five attributes (i.e., type, speed, previous speed, In the Evaluate Experimental Results step, the performance
GMTI performance, and position) in the previous section, of estimation and prediction for the learned reasoning model is
become five resident nodes in a candidate MTheory using assessed by the performance criteria in the requirements
MEBN-RM. For example, the type attribute for the target g defined in the Analyze Requirement step (e.g., a sum of KL
becomes a resident node Type(g). The speed attribute for the divergence values < 0.001). If the measured value satisfies the
target g and at the time t becomes a resident node Speed(g, t). criteria, the learned reasoning model is accepted and this step
The two rules enable the candidate MTheory to have results in the learned reasoning model. If the requirement is not
conditional dependencies between the resident nodes. For satisfied, we can return to the previous steps to improve the
example, from Rule 1, a conditional probability of the resident performance of the learned reasoning model.
node Speed(g, t) given the resident node Type(g) can be
derived. IV. CASE STUDY
In this section, we introduce a case study for the HMLP
The Learn Reasoning Model step contains MTheory process. In the case study, we aimed to develop a PSAW
learning from relational data. MTheory learning is to learn an MTheory for a prototype software PSAW system, called
MTheory M* given a training dataset D in RM and an initial HERALD, which wards off attacks against critical
MTheory M. MTheory learning can be written as follows: infrastructure by means of early detection of threatening targets,
identification of the targets, estimation of the target’s activities,
𝑀𝐸𝐵𝑁!"!!"#$_!"#$%&%' ( 𝑀 ∗ ∶ 𝑫, 𝑀 ) (1) and prediction of virtual short-term future situations. A
decision maker in charge of critical infrastructure (CI)
MTheory learning can use a score based learning method in protection must examine situations, analyze the situations to
which candidate parameters and structures for the MTheory are determine crucial aspects, choose courses of action to address
searched and a score function [18] is used to evaluate the the mission, and execute the decisions. The goal of HERALD
theories given the dataset. If we use score based MEBN was to support some of these activities for the decision maker
learning, the score function must be defined [28][29][30][31]. by fusing sensor system reports. The sensor systems producing
Search algorithms for structures and parameters of the model the reports were of various types such as a Moving Target
are chosen. The results of the search are scored according to Indicator (MTI), an Imagery Intelligence (IMINT) system, and
the score function. A high scoring reasoning model is selected a Geospatial Intelligence (GEOINT) system. A PSAW
as an output in this step. MTheory developer in charge of the development of a
HERALD MTheory cooperated with the decision maker to
D. Test Reasoning Model
develop the HERALD MTheory by using the HMLP process.
In the Test Reasoning Model step, a learned reasoning The following shows how the development operates.
model is evaluated to determine whether to accept it. The
accepted reasoning model is output as a final result in this step. A. Analyze Requirements
This step is decomposed into two sub-steps (Fig. 6): (1) an HERALD focused on threats from terrorists and UAVs
Experiment Reasoning Model step and (2) an Evaluate approaching the critical infrastructure element. Operational
Experimental Results step. scenarios for attacker (Red Team) and defender (Blue Team)
are given below.
(1) Red Team Scenarios
HERALD considered two Red Team scenarios attacking a
nuclear power plant in South Korea: (1) terrorist attack and (2)
UAV attack. Both scenarios are performed in the same
environmental conditions in which the weather is clear, it is
evening, and the season is the end of the spring. (1) In the
Fig. 6. Test Reasoning Model terrorist attack scenario, a UH-60 helicopter carrying three
armed terrorists heads toward the nuclear power plant. At 800
The Experiment Reasoning Model step tests the learned m from the region, the helicopter airdrops the terrorists. The
reasoning model using a test dataset. The test dataset can be terrorists break through a surveillance system and infiltrate the
generated from simulations, existing data and/or actual plant. The terrorists occupy a critical infrastructure element at
experiments. This experiment can consist of the following five the plant, set up explosives at a vulnerable point of the critical
steps. (1) The learned reasoning model is exercised on a test infrastructure element, and evacuate from the area. (2) In the
case from the test dataset. (2) The test dataset provides ground UAV attack scenario, three UAVs flying in formation at speed
truth data for a certain metric (e.g., KL divergence) in the 74MPH head toward a nuclear power plant. Each UAV
requirements defined in the Analyze Requirement step. (3) The contains a bomb. The UAVs move to 12 miles from the plant
metric is used to measure performance between results from and scatter to distract the Blue Team from surveillance. At 6
the learned reasoning model and the ground truth data. (4) miles from the plant, each UAV flies into the plant and
Steps 1-3 are repeated for all testing cases. (5) This step results detonates its bomb.
in a result value integrating all measured values. Any attribute (2) Blue Team Scenarios
for which we have no ground truth data are ignored in this
process. For the two Red Team scenarios, the Blue Team responds to
the Red Team. A Blue Team command center uses sensors
around the plant for PSAW. If targets are identified, the B. Design World Model and Rules
command center reacts to the targets (e.g., double-check using This step contains two sub-steps (Design World Model and
a mobile sensor). If the targets are identified more precisely, Design Rules). The first sub-step is to define a world model for
the command center defends the operations from the targets. HERALD from the requirements, domain knowledge and/or
The Blue Team scenarios vary according to discrimination for a existing data schemas (e.g., RM). The reference MEBN model
target. The target is differently identified by its location. For for PSAW can be used in this step to identify possible entities
example, if a target is located far from a sensor, the sensor may and variables (or attributes in RM). The reference model
judge whether the target is detected or not. This case means provides information about possible entities (i.e., T, OR, SR,
that the target is in a detection layer. Target discrimination is a TR, and IT). Entities derived from these categories can have
process in which the target is assigned to an object type from a various attributes and be related to other entities. Fig. 7 shows
set of object types. Target discrimination can be broken into an EER model for HERALD as the HERALD world model.
four categories: (1) Detection, (2) Classification, (3)
Recognition, and (4) Identification, or DCRI [27]. In the
detection layer, an unconfirmed object has been acquired (e.g.,
object of interest). In the classification layer, the unconfirmed
object is determined by class (e.g., vehicle or animal). In the
recognition layer, threatening targets are determined (e.g., tank
or armored car). In the identification layer, the specific model
of the threatening target is determined (e.g., M1A2 or AH-
64D). The scenario of the Blue Team is as follows. (1) If a
target enters the detection layer, the commander of the Blue
Team observes the target’s activity closely. The target can
either enter the classification layer or leave the detection layer.
(2) If the target enters the classification layer, the commander
of the Blue Team tries to identify whether the target is armed or
not. The target can either enter the recognition layer or leave
the classification layer. (3) If the target enters in the
recognition layer and is determined as a threatening target, the
commander of the Blue Team prepares against an attack from
the target. The target can move to some places, stay at a place,
prepare for an attack, or attack the critical infrastructure. (4) If
the target enters the identification layer, the commander of the
Blue Team orders attack against the target. The target engages Fig. 7. Part of EER Model for HERALD
with the Blue team.
(3) Requirements The EER model for HERALD contained 15 relations (i.e.,
sensor, target, time, ci_situation, reportedtarget_mtirpt,
From the operational scenarios for the Red Team and the
sensorof, reportedtarget_imintsrpt, reportedtarget_geointsrpt,
Blue Team, we developed goals, queries and evidence for the
sensortemporalproperty, targettemporalproperty, predecessor,
PSAW questions Of Section 3. Two examples are given:
criticalinfrastructure, mti_report, and imints_report,
Goal 1: Recognize emergency situation for a critical geoints_report). For example, the target relation possesses a
infrastructure primary key TargetID and three attributes TargetType,
TargetSize, and TargetImage. We grouped relations and
Query 1.1: How high is the level of danger to the attributes in the EER model for HERALD using the reference
critical infrastructure? MEBN model for PSAW. The reference model suggested four
Evidence 1.1.1: All reports for targets groups: (1) Relation of Observed Target, (2) Sensing
Conditions, (3) Relation of Interpreted Target, and (4)
Goal 2: Identify threatening target Situation. The Relation of Observed Target group contained
Query 2.1: What is the type of the target? the relations target and targettemporalproperty. The Sensing
Conditions group contained the relation
Evidence 2.1.1: All reports for targets sensortemporalproperty. The Relation of Interpreted Target
For our case study, we used the Brier score for a multi- group contained the relaions mti_report, and imints_report,
category forecast, where perfect prediction yields a score of geoints_report. The reference MEBN model provided some
zero, and a sure prediction on the wrong hypothesis yields a rule information between these groups. The Sensing Conditions
score of two [17], as a metric for a performance criterion. The group and the Relation of Observed Target group influenced
performance criterion was that the Brier score for reasoning the Relation of Interpreted Target group. For example, the
results (e.g., results for Query 2.1) from the HERALD attribute DistanceToSensor in the relation sensortempora-
MTheory should be less than 0.5. lproperty influenced the report attributes in the report relations
mti_report, and imints_report, geoints_report. The arrows in
Fig. 7 indicate these relationships. From the relationships and
domain knowledge, we derive the following rules.
Rule 1: causal({TargetType, Mission}, DangerLevel} we used parameter learning given the candidate HERALD
MTheory. The training dataset was generated by the HERALD
Rule 2: causal(TargetType, {TargetSize, TargetImage , simulator. The HERALD simulator simulated ground truth
Temperature})
information about a situation in which blue and red teams
Rule 3: causal(Temperature, TemperatureReport) operated against each other. The Red Team could be a group of
terrorists or UAVs (unmanned aerial vehicles). The Red Team
C. Construct Reasoning Model aimed to destroy a critical infrastructure element defended by
This step consists of two sub-steps (Map to Reasoning the Blue Team. The Blue Team had several types of sensors
Model and Learn Reasoning Model) to construct a reasoning that could detect threatening targets of interest, and a command
model from a training dataset. According to MEBN-RM, an center, charged with using data from the sensors to identify the
entity relation which contains only one attribute for the primary situation and make decisions about future actions. The
key of the relation (e.g., sensor and target) can be defined as an simulator simulated the operational scenarios in Section 4. To
entity type in MEBN. Thus, there were 7 entity types (sensor, learn the HERALD MTheory, the training dataset was
target, time, criticalinfrastructure, reportedtarget_mtirpt, generated from the HERALD simulator. A test dataset was also
reportedtarget_imintsrpt, and reportedtarget_geointsrpt). Also, generated from the HERALD simulator to be used for
each of the attributes in the relations could be mapped to a evaluation. These datasets were separately stored in a relational
resident node in MEBN using MEBN-RM. For example, the database.
attribute Activity of the relation targettemporalproperty
became the resident node Activity(tr, t). The attribute D. Test Reasoning Model
DangerLevel of the relation ci_situation became the resident This step consists of two sub-steps (Experiment Reasoning
node DangerLevel(ci, t). Model and Evaluate Experimental Results) to evaluate a
reasoning model from the test dataset. In the Experiment
Rules which are defined in the previous step are used to Reasoning Model step, the performance of estimation and
develop conditional dependence and independence between prediction for the learned HERALD MTheory was assessed
resident nodes. From Rule 1, we had a conditional dependence using some measures (e.g., accuracy, sensitivity, specificity,
P(DangerLevel | TargetType, Mission). This relationship could AUC, and Brier score). This experiment consisted of the
be represented in a candidate MTheory for HERALD. following five steps. (1) The test dataset provided entity
information (e.g., target1, sensor1, and time1) and simulated
ground truth information (e.g., ImageReport = Unknown,
TemperatureReport = High, SizeReport = Small) to the learned
HEARLD MTheory. (2) Given these, a marginal probability of
a query (e.g., Query 2.1: P(TargetType | ImageReport =
Unknown, TemperatureReport = Mid, SizeReport = Small) was
calculated using the learned HEARLD MTheory. (3) The test
dataset provided a ground truth data (e.g., TargetType =
ThreateningTarget). (4) The steps 1-3 were repeated for all
Fig. 8. CI Situation MFrag in the HERALD MTheory testing cases. (5) Finally, for results for all cases, the measures
were calculated.
Fig. 8 shows an MFrag, called a CI Situation MFrag in the
Accuracy Sensitivity Specificity AUC Brier
candidate HERALD MTheory. Note that this MFrag is a part
of the HERALD MTheory, which contains several MFrags Average 0.5893 0.9664 0.5884 0.6384 0.0294
(e.g., MTI_Report, IMINTS_Report, and GEOINTS_Report) for Standard
higher level information fusion. The MFrag represents 0.0049 0.0434 0.0049 0.0078 0.0008
Deviation
probabilistic knowledge of critical infrastructure situations Table 1. Evaluation results for target type estimation
involving a group of targets. It contains three isA context nodes
for entities Target, CriticalInfrastructure, and Time, two input In the Evaluate Experimental Results step, we evaluated the
nodes TargetType (with states Others and ThreateningTarget) Brier score. Table 1 shows the accuracy, sensitivity, specificity,
and Mission (with states Others and Attacking), and one AUC, and Brier score for the target type estimation, in which
resident node DangerLevel (with states Dangerous and Safe). the four test datasets were used to evaluate the learned
The resident node DangerLevel(ci, t) has a critical MTheory. Each test dataset contained 10000 cases. For the
infrastructure ordinary variable ci and a time ordinary variable target type estimation, the average and the standard deviation
t as arguments. It indicates whether the situation of the group of the Brier score were 0.0294 and 0.0008, respectively. This
of targets in the times t is dangerous or safe. The resident node result satisfied the requirement defined in the Analyze
DangerLevel(ci, t) is influenced by the input node Requirements step.
TargetType(tr) with a target ordinary variable tr and Mission(tr,
t) with the target ordinary variable tr and the time ordinary V. CONCLUSION
variable t. We have developed MEBN models in PSAW.
Developments for PSAW-MEBN models were labor-intensive
In the Learn Reasoning Model step, the candidate and insufficiently agile processes. To overcome these problems,
HERALD MTheory is refined using a MEBN learning MEBN machine learning for the PSAW-MEBN models was
algorithm from a training dataset. In our simple learning test, developed. In the real world, limitations on available data may
necessitate using expert judgment. For this reason, a process [16] Laskey, K. B., & Mahoney, S. M. (2000). Network engineering for agile
for Human-aided MEBN learning in PSAW was presented in belief network models. Knowledge and Data Engineering, IEEE
Transactions on, 12(4), 487-498.
this paper. The process relied partially on expert's knowledge
[17] Brier, G. W. (1950). Verification of forecasts expressed in terms of
and insight. One case study was conducted for the process. A probability. Monthly weather review, 78(1), 1-3.
full effectiveness analysis to evaluate improvement in the [18] Park, C. Y., Laskey, K. B., Costa, P. C. G., & Matsumoto, S. (2013).
quality of the reasoning and reduction in time cost and time for Multi-Entity Bayesian Networks Learning For Hybrid Variables In
development was beyond the scope of this paper, and is left for Situation Awareness. Proceedings of the 16th International Conference
future research. Our future plans are to elaborate the process on Information Fusion (Fusion 2013).
and to apply the process to more realistic scenarios. [19] Park, C. Y., Laskey, K. B., Costa, P. C. G., & Matsumoto, S. (2013).
Multi-Entity Bayesian Networks Learning In Predictive Situation
Awareness. Proceedings of the 18th International Command and Control
Technology and Research Symposium (ICCRTS 2013).
ACKNOWLEDGMENTS [20] Waltz, E. L., & Buede, D. M. (1986). Data fusion and decision support
The research was partially supported by the Office of Naval for command and control. IEEE Transactions on Systems, Man, and
Cybernetics, 6(16), 865-879.
Research (ONR), under Contract#: N00173-09-C-4008.
[21] Dorion, É., Gareau, M., & Roy, J. (2008). Situation Analysis for the
Tactical Army Commander (SATAC).
REFERENCES
[22] Roy, J. (2001, August). From data fusion to situation analysis. In
Proceedings of Fourth International Conference on Information Fusion
[1] Laskey, K. B. (2008). MEBN: A Language for First-Order Bayesian (Vol. 2).
Knowledge Bases. Artificial Intelligence, 172(2-3).
[23] Codd, E. F. (1970). A Relational Model of Data for Large Shared Data
[2] Laskey, K. B., D’Ambrosio, B., Levitt, T. S., & Mahoney, S. M. (2000). Banks. Communications of the ACM.
Limited Rationality in Action: Decision Support for Military Situation
[24] Codd, E. F. (1969). Derivability, Redundancy, and Consistency of
Assessment. Minds and Machines, 10(1), 53-77.
Relations Stored in Large Data Banks. IBM Research Report.
[3] Wright, E., Mahoney, S. M., Laskey, K. B., Takikawa, M. & Levitt, T.
[25] Carvalho, R. N. (2011). Probabilistic Ontology: Representation and
(2002). Multi-Entity Bayesian NetworksforSituation Assessment.
Modeling Methodology. PhD Dissertation. George Mason University.
Proceedings of the Fifth International Conference on Information Fusion.
[26] Kullback, S., & Leibler, R. A. (1951). On information and sufficiency.
[4] Costa, P. C. G., Laskey, K. B., Takikawa, M., Pool, M., Fung, F., &
The annals of mathematical statistics, 79-86.
Wright, E. J. (2005). MEBN Logic: A Key Enabler for Network Centric
Warfare. Proceedings of the 10th ICCRTS. [27] Self, M., Miller, B., & Dixon, D. (2005). Acquisition Level Definitions
and Observables for Human Targets Urban Operations and the Global
[5] Suzic, R. (2005). A generic model of tactical plan recognition for threat
War on Terrorism. Army communications-electronics command fort
assessment. In Defense and Security (pp. 105-116). International Society
belvoir va night vision and electronics sensors directorate.
for Optics and Photonics.
[28] Cooper, G. F. & Herskovits, E. (1992). A Bayesian method for the
[6] Costa, P. C. G., Laskey, K. B., & Chang, K. C. (2009). PROGNOS:
induction of probabilistic networks from data. Machine Learning 9(4),
Applying Probabilistic Ontologies To Distributed Predictive Situation
309–347.
Assessment In Naval Operations. Proceedings of the 14th ICCRTS.
[29] Bouckaert, R. R. (1993). Probabilistic network construction using the
[7] Carvalho, R. N., Costa, P. C. G., Laskey, K. B., & Chang, K. C. (2010).
minimum description length principle. In Symbolic and Quantitative
PROGNOS: predictive situational awareness with probabilistic
Approaches to Reasoning and Uncertainty: European Conference
ontologies. In Proceedings of the 13th International Conference on
ECSQARU ’93, Lecture Notes in Computer Science 747, 41–48,
Information Fusion.
Springer.
[8] Pearl, J. (1988). Probabilistic Reasoning in Intelligent Systems:
[30] Heckerman, D., Geiger, D., & Chickering, D. M. (1995). Learning
Networks of Plausible Inference. San Mateo, CA, USA: Morgan
Bayesian networks: The combination of knowledge and statistical data.
Kaufmann Publishers.
Machine Learning, 20:197–243.
[9] Laskey, K., Alghamdi, G., Wang, X., Barbara, D., Shackelford, T.,
[31] Chickering, D. M., Geiger, D. & Heckerman, D. (1996). Learning
Wright, E., & Fitzgerald, J. (2004). In Proceedings of the Conference on
Bayesian networks: search methods and experimental results. In
Behavioral Representation in Modeling and Simulation.
Learning from Data: Artificial Intelligence and Statistics V, Fisher, D. &
[10] Endsley, M. R. (1988). Design and evaluation for situation awareness Lenz, H.-J. (eds). Lecture Notes in Statistics 112, 112–128. Springer.
enhancement. Paper presented at the Human Factors Society 32nd
[32] Haberlin Jr, R. J., da Costa, P. C., & Laskey, K. B. (2013). A Reference
Annual Meeting, Santa Monica, CA.
Architecture for Probabilistic Ontology Development. STIDS 2013.
[11] Endsley, M. R., Bolte, B., & Jones, D. G. (2003). Designing for
situation awareness: An approach to human-centered design. New York,
NY: Talyor & Francis.
[12] Breton, R., & Reousseau, R. (2001). Situation Awareness: A Review of
the Concept and its Measurement.Technical Report No. 2001-220,
Defense Research and Development Canada, Valcartier.
[13] Koller, D., & Pfeffer, A. (1997). Object-Oriented Bayesian Networks.
Paper presented at the Thirteenth Conference on Uncertainty in
Artificial Intelligence (UAI-97). SanFrancisco, CA, USA.
[14] Friedman, N., Getoor, L., Koller, D., & Pfeffer, A. (1999). Learning
Probabilistic Relational Models. In Proceedings of the Sixteenth
International Joint Conference on Artificial Intelligence (IJCAI-99).
[15] Park, C. Y., Laskey, K. B., Costa, P. C. G., & Matsumoto, S. (2014).
Predictive Situation Awareness Reference Model using Multi-Entity
Bayesian Networks. Proceedings of the 17th International Conference on
Information Fusion.

View publication stats

Vous aimerez peut-être aussi