Académique Documents
Professionnel Documents
Culture Documents
logic
programming
MDL
principle
Bayesian
networks
neural
networks
probabilistic
Horn
abduction
possible
worlds
semantics
logical
abduction
Overview
Causal and evidential modelling & reasoning
Bayesian networks, Bayesian conditioning &
abduction
Noise, overfitting, and Bayesian learning
vision:
effects
of interest
observed
scene image
model data
4
Evidential modelling:
effects causes
vision:
image scene
data model
cause
causal reasoning
prediction
6
Bayes Rule
[de Moivre 1718, Bayes 1763, Laplace 1774]
P(e|h)P(h)
P(h|e) =
P(e)
Proof:
P(h e) = P(e|h)P(h)
= P(h|e)P(e)
Lesson #1
You should know the difference between
evidential & causal modelling
evidential & causal reasoning
There seems to be a relationship between Bayes
theorem and abduction used for the same task.
Overview
Causal and evidential modelling & reasoning
Bayesian networks, Bayesian conditioning &
abduction
Noise, overfitting, and Bayesian learning
10
Bayesian Networks
Graphical representation of independence.
DAGs with nodes representing random
variables.
Embed independence assumption:
If b1 , , bn are the parents of a then
P(a|b1 , , bn , V ) = P(a|b1 , , bn )
if V is not a descendant of a.
11
light_switch_on
projector_plugged_in
projector_switch_on
power_in_wire
power_in_projector
room_light_on
lamp_works
projector_lamp_on
alan_reading_book
mirror_working
screen_lit_up
ray_is_awake
ray_says_"screen is dark"
12
13
Probabilities of hypotheses
P(projector_working_ok)
= P(projector_lamp_on |
power_in_projector lamp_works)
provided as part of Bayesian network
P(projector_working_ok)
= 1 P(projector_working_ok)
14
P(g) =
P(E)
E is an explanation of g
P(E) =
Y
hE
P(h)
given with logic program
16
Conditional Probabilities
P(g e) explain g e
P(g|e) =
P(e) explain e
Given evidence e, explain e, then try to explain g
from these explanations.
17
Lessons #2
Bayesian conditioning = abduction
The evidence of Bayesian conditioning is what
is to be explained.
Condition on all information obtained since
the knowledge base was built.
P(h|e k) Pk (h|e)
18
Overview
Causal and evidential modelling & reasoning
Bayesian networks, Bayesian conditioning &
abduction
Noise, overfitting, and Bayesian learning
19
Induction
data
evidential reasoning
model
causal reasoning
prediction
20
Potential Confusion
Often the data is about some evidential reasoning
task. e.g., classification, diagnosis, recognition
Example: We can do Bayesian learning where the
hypotheses are decision trees, neural networks,
parametrized distribution, or Bayesian networks.
Example: We can do hill climbing learning with
cross validation where the hypotheses are decision
trees, neural networks, parametrized distribution,
or Bayesian networks.
21
fit to data
bias
P (e|h)P (h)
P (h|e) =
P (e)
normalizing constant
23
P(e|h)P(h)
= arg max
h
P(e)
= arg max P(e|h)P(h)
h
= arg max
h
log2 P(e|h)
log2 P(h)
z
}|
{
z
}|
{
the number of bits to
the number of
describe the data in + bits to describe
terms of the model
the model
24
heads
...
headsN
25
Explaining Data
If you observe:
heads(c1 ), tails(c2 ), tails(c3 ), heads(c4 ), heads(c5 ), . . .
For each [0, 1] there is an explanation:
{prob_heads( ), turns_heads(c1 , ), turns_tails(c2 , ),
turns_tails(c3 , ), turns_heads(c4 , ), turns_heads(c5 , ),
. . .}
27
h1
h2
o1
o2
o3
prop(X, o2 , V )
prop(X, h1 , V1 ) prop(X, h2 , V2 )
abducible
param(p8 , P1 ) param(p10 , P2 )
1
.
sigmoid additive
V=
(V
P
+V
P
)
1+e 1 1 2 2
28
Lesson #3
Abductive view of Bayesian learning:
rules imply data from parameters or possible
representations
parameter values or possible representations
abducible
rules contain logical variables for data items
29
Neural Nets
Bayesian Nets
evidential
causal
sigmoid additive
linear gaussian
abduction
causal reasoning
none
direct
context changes
fragile
robust
learning
easier
more difficult
30
Conclusion
Bayesian reasoning is abduction.
Bayesian network is a causal modelling language
abduction + probability.
What logic-based abduction can learn from Bayesians:
Handle noise, overfitting
Conditioning: explain everything
Algorithms: exploit sparse structure, exploit extreme
distributions, or stochastic simulation.
What the Bayesians can learn:
Richer representations.
31