Académique Documents
Professionnel Documents
Culture Documents
Knowledge representation
Joint probability distribution
can answer any question about the domain can become intractably large as #RV grows can be difficult to specify P for atomic events
Conditional independence can simplify P assign t A data structure - a belief network or Bayesian network that represents the dependence between variables and gives a concise specification of the joint.
CSE 471/598, CBS 598 by H. Liu 2
A set of random variables A set of directed links connects pairs of nodes Each node has a conditional P table that quantifies the effects that the parents have on the node The graph has no directed cycles (DAG)
It is usually much easier for an expert to decide conditional dependence relationships than specifying probabilities Sometimes, experts can have very different opinions
CSE 471/598, CBS 598 by H. Liu 3
Once the network is specified, we need only specify conditional probabilities for the nodes that participate in direct dependencies, and use those to compute any other probabilities. A simple Bayesian network (Fig 14.1) An example of burglary-alarm-call (Fig 14.2) The topology of the network can be thought of as the general structure of the causal process. Many details (Mary listening to loud music, or phone ringing and confusing John) are summarized in the uncertainty associated with the links from Alarm to JohnCalls and MaryCalls.
CSE 471/598, CBS 598 by H. Liu 4
The probabilities actually summarize a potentially infinite set of possible circumstances Specifying the CPT for each node (Fig 14.2)
A conditioning case - a possible combination of values for the parent nodes (2^n) Each row in a CPT must sum to 1 A node with no parents has only one row (priors)
Representing the JPD - helpful in understanding how to construct networks Representing conditional independence relations helpful in designing inference procedures
P(x1, ,xn)=4P(xi|Parents(xi))
What s the probability of the event of J^M^A^!B^!E?
(14.1)
Pick a var and add a node to the network Set its parents to some minimal set of nodes already in the net to satisfy Eq.14.2 Define the CPT for the var.
CSE 471/598, CBS 598 by H. Liu 9
Compactness
A Bayesian network can often be far more compact than the full joint. In a locally structured system, each subcomponent interacts directly with only a bounded number of other components. Local structure is usually associated with linear rather than exponential growth in complexity. With 30 (n) nodes, if a node is directly influenced by 5 (k) nodes, what s the difference between BN & joint?
Node ordering
The correct order to add nodes is to add the root causes first, then the variables they influence, and so on until we reach the leaves that have no direct causal influence on the other variables. What happens if we happen to choose the wrong order? Fig 14.3 shows an example. If we stick to a true causal model, we end up having to specify fewer numbers, and the numbers will often be easier to come up with.
CSE 471/598, CBS 598 by H. Liu 11
A node is cond independent of all other nodes in the network, given its parents, children, and children s parents (its Markov blanket).
Representation of CPTs
Given canonical distributions, the complete table can be specified by naming the distribution with some parameters. A deterministic node has its value specified exactly by the values of its parents. Uncertain relationships can often be characterized by noisy logical relationships. An example for determine cond probabilities starting with P(!Fever) on page 501 given the individual inhibition probabilities as
Inference by enumeration The variable elimination algorithm The complexity of exact inference Clustering algorithms Direct sampling methods
Rejection sampling Likelihood weighting
Approximate inference
14
15
Strict logic reasoning (ignore uncertainty) Probabilistic techniques using the full Joint Default reasoning - believed until a better reason is found to believe something else Rules with certainty factors Handling ignorance - Dempster-Shafer theory Vagueness - something is sort of true (fuzzy logic)
Probability makes the same ontological commitment as logic: the event is true or false
CSE 471/598, CBS 598 by H. Liu 16
Default reasoning
The four-wheel car conclusion is reached by default. New evidence can cause the conclusion retracted, while FOL is strictly monotonic. Representatives are default logic, nonmonotonic logic, circumscription There are problematic issues
Details in Chapter 10
17
Rule-based methods
Logical reasoning systems have properties like:
These properties are good for obvious computational advantages; bad as they re inappropriate for uncertain reasoning.
CSE 471/598, CBS 598 by H. Liu 18
Summary
Reasoning properly
In FOL, it means conclusions follow from premises In probability, it means having beliefs that allow an agent to act rationally
Conditional independence info is vital A Bayesian network is a complete representation for the JPD, but exponentially smaller in size Bayesian networks can reason causally, diagnostically, intercausally, or combining two or more of the three. For polytrees, the computational time is linear in network size.
CSE 471/598, CBS 598 by H. Liu 19