SlideShare a Scribd company logo
1 of 38
Download to read offline
Artificial Intelligence Reasoning with Uncertainty
Uncertainty
Reading: Section 13.1 - 13.6 of Textbook R&N
In which we see what an agent should do when not all is
crystal clear.
By Bal Krishna Subedi 1
Artificial Intelligence Reasoning with Uncertainty
Uncertain Knowledge
Let action At = leave for airport t minutes before flight
Will At get me there on time?
Problems:
1) partial observability (road state, other drivers' plans, etc.)
2) noisy sensors (radio traffic reports)
3) uncertainty in action outcomes (flat tyre, etc.)
4) complexity of modeling and predicting traffic
Hence a purely logical approach either
1) risks falsehood: “A25 will get me there on time” or
2) leads to conclusions that are too weak for decision making: “A25 will get me
there on time if there's no accident on the bridge and it doesn't rain and my
tires remain intact etc etc.”
A1440 might reasonably be said to get me there on time but I'd have to stay
overnight in the airport…
By Bal Krishna Subedi 2
Artificial Intelligence Reasoning with Uncertainty
Handling Uncertainty
Instead of providing all condition it can express with degree of beliefs in the relevant
sentences.
Example:
Say we have a rule
if toothache then problem is cavity
But not all patients have toothaches because of cavities (although perhaps most do)
So we could set up rules like
if toothache and not(gum disease) and not(filling) and ......then problem is cavity
This gets very complicated! a better method would be to say
if toothache then problem is cavity with probability 0.8
Given the available evidence,
A25 will get me there on time with probability 0.04
A most important tool for dealing with degree of beliefs is probability theory, which
assigns to each sentence a numerical degree of belief between 0 & 1.
By Bal Krishna Subedi 3
Artificial Intelligence Reasoning with Uncertainty
Making decisions under uncertainty
Suppose I believe the following:
P(A25 gets me there on time|…) = 0.04
P(A90 gets me there on time|…) = 0.70
P(A120 gets me there on time|…) = 0.95
P(A1440 gets me there on time|…) = 0.9999
Which action to choose?
Depends on my preferences for missing flight vs. length of wait at airport, etc.
Utility theory is used to represent and infer preferences
Decision theory = utility theory + probability theory
The rational decision depends on both the relative importance of various goals
and the likelihood that, and degree to which, they will be achieved.
By Bal Krishna Subedi 4
Artificial Intelligence Reasoning with Uncertainty
Probability basics
Begin with a set Ω - the sample space
e.g., 6 possible rolls of a die.
ω ε Ω is a sample point/possible world/atomic event
A probability space or probability model is a sample space with
an assignment P(ω) for every ω ε Ω s.t.
0 ≤ P(ω) ≥ 1
∑P(ω) = 1
e.g., P(1) = P(2) = P(3) = P(4) = P(5) = P(6) = 1/6.
An event A is any subset of Ω
P(A) = ∑{ωεA} P(ω)
E.g., P(die roll < 4) = 1/6 + 1/6 + 1/6 = 1/2
By Bal Krishna Subedi 5
Artificial Intelligence Reasoning with Uncertainty
Random variables
By Bal Krishna Subedi 6
An application domain (a world) is described by a set of random variables, each
variable referring to a part of the domain whose status is initially unknown.
Example:
A simple world consisting of two random variables:
Cavity – a Boolean variable that refers to whether my lower left wisdom
tooth has a cavity
Toothache - a Boolean variable that refers to whether I have a toothache or
not.
We use the single capital letters to represent unknown random variables
P induces a probability distribution for any r.v. X:
P(X = xi) = ∑ {ω: X(ω) = xi}P(ω)
e.g., P(Odd = true) = 1/6 + 1/6 + 1/6 = 1/2
Each RV has a domain of values that it can take it.
e. g. domain of Cavity is {true, false}
RVs domain are: Boolean, Discrete and Continiuous
Artificial Intelligence Reasoning with Uncertainty
Atomic events
An atomic event is a complete specification of the state of
the world about which the agent is uncertain.
Example:
In the above world with two random variables (Cavity and
Toothache) there are only four distinct atomic events, one
being:
Cavity = false, Toothache = true
Which are the other three atomic events?
By Bal Krishna Subedi 7
Artificial Intelligence Reasoning with Uncertainty
Propositions
Think of a proposition as the event (set of sample points) where the proposition is
true
Given Boolean random variables A and B:
event α = set of sample points where A(ω) = true
event ¬α = set of sample points where A(ω) = false
event a ^ b = points where A(ω) = true and B(ω) = true
Often in AI applications, the sample points are defined by the values of a set of
random variables, i.e., the sample space is the Cartesian product of the ranges of
the variables.
With Boolean variables, sample point = propositional logic model
e.g., A = true, B = false, or a ^ ¬b.
Proposition = disjunction of atomic events in which it is true
e.g., (a ∨ b) ≡ (¬a ^ b) ∨ (a ^ ¬b) ∨ (a ^ b)
P(a ∨ b) = P(¬a ^ b) + P(a ^ ¬b) + P(a ^ b)
By Bal Krishna Subedi 8
Artificial Intelligence Reasoning with Uncertainty
Syntax for propositions
Propositional or Boolean random variables
e.g., Cavity (do I have a cavity?)
Discrete random variables (finite or infinite)
e.g., Weather is one of (sunny, rain, cloudy, snow)
Weather = rain is a proposition
Values must be exhaustive and mutually exclusive
Continuous random variables (bounded or unbounded)
e.g., Temp = 21.6, also allow, e.g., Temp < 22.0.
By Bal Krishna Subedi 9
Artificial Intelligence Reasoning with Uncertainty
Prior probability
The prior or unconditional probability associated with a proposition is the degree
of belief accorded to it in the absence of any other information.
Example:
P(Weather = sunny) = 0.72
P(Cavity = true) = 0.1 or P(cavity) = 0.1
Probability distribution gives values for all possible assignments:
P(Weather) = (0.72, 0.1, 0.08, 0.1)
Joint probability distribution for a set of r.v.s gives the probability of every
atomic event on those r.v.s (i.e., every sample point)
P(Weather, Cavity) = a 4 × 2 matrix of values.
By Bal Krishna Subedi 10
Every question about a domain can be answered by the joint distribution
because every event is a sum of sample points
Artificial Intelligence Reasoning with Uncertainty
Conditional probability
The conditional probability “P(a|b)” is the probability of “a” given that all
we know is “b”.
e.g., P(cavity|toothache) = 0.8 means if a patient have toothache and no other
information is yet available, then the probability of patient’s having the cavity is
0.8
Definition of conditional probability:
P(a|b) = P(a ^ b)/P(b) if P(b) ≠ 0
Product rule gives an alternative formulation:
P(a ^ b) = P(a|b)P(b) = p(b|a)P(a)
Chain rule is derived by successive application of product rule:
P(X1,…,Xn) = P(X1,…,Xn-1) P(Xn|X1,…,Xn-1)
……..
= ∏n
i=1P(Xi|X1,…,Xi-1)
By Bal Krishna Subedi 11
Artificial Intelligence Reasoning with Uncertainty
The axioms of probability
1. All probabilities are between 0 and 1:
0 ≤ P(a) ≤ 1
2. Necessarily true propositions have probability 1 and
necessarily false propositions have probability 0:
P(true) = 1 P(false) = 0
3. The probability of a disjunction is given by:
P(a U b) = P(a) + P(b) – P(a ∩ b)
By Bal Krishna Subedi 12
Inference using full joint probability
distribution
Artificial Intelligence Reasoning with Uncertainty
We use the full joint distribution as the knowledge base
from which answers to all questions may be derived.
The probability of a proposition is equal to the sum of the
probabilities of the atomic events in which it holds.
P(a) = Σ P(ei)
Therefore, given a full joint distribution that specifies the
probabilities of all the atomic events, one can compute the
probability of any proposition.
By Bal Krishna Subedi 13
Artificial Intelligence Reasoning with Uncertainty
Full joint probability distribution: example
We consider the following domain consisting of three Boolean variables:
Toothache, Cavity, and Catch (the dentist’s nasty steel probe catches in my tooth).
The full joint distribution is the following 2x2x2 table:
toothache ¬toothache
cavity
¬cavity
catch ¬catch catch ¬catch
0.108 0.012 0.072 0.008
0.016 0.064 0.144 0.576
The probabilities in the joint distribution must sum to 1.
The probability of any proposition can be computed from the probabilities in
the table.
By Bal Krishna Subedi 14
Each cell represents an atomic event and these are all the possible atomic
events.
Artificial Intelligence Reasoning with Uncertainty
Full joint probability distribution: example
toothache ¬toothache
cavity
¬cavity
catch ¬catch catch ¬catch
0.108 0.012 0.072 0.008
0.016 0.064 0.144 0.576
We simply identify those atomic events in which the proposition is true and add up
their probabilities
P(cavity or toothache) = P(cavity, toothache, catch) + P(cavity, toothache,
¬catch) + P(cavity, ¬toothache, catch) +
P(cavity, ¬toothache, ¬catch) + P(¬cavity,
toothache, catch) + P(¬cavity, toothache, ¬catch)
= 0.108+0.012+0.072+0.008+0.016+0.064=0.28
By Bal Krishna Subedi 15
Artificial Intelligence Reasoning with Uncertainty
Marginalization or summing out
toothache ¬toothache
cavity
¬cavity
catch ¬catch catch ¬catch
0.108 0.012 0.072 0.008
0.016 0.064 0.144 0.576
P(cavity)
P(¬toothache)
P(cavity, ¬toothache)
That is, a distribution over Y can be obtained by summing out all
the other variables from any joint distribution containing Y.
P(Y) = Σ P(Y, z)
P(cavity) = 0.108+0.012+0.072+0.008 = 0.2
P(¬toothache) = 0.072+0.008+0.144+0.576 = 0.8
P(cavity, ¬toothache) = 0.072+0.008 = 0.08
By Bal Krishna Subedi 16
Artificial Intelligence Reasoning with Uncertainty
Conditioning
P(Y) = Σ P(Y, z)
P(Y, z) = P(Y|z)P(z)
Therefore, for any set of variables Y and Z:
P(Y) = Σ P(Y|z)P(z) - This is the conditioning rule
toothache ¬toothache
cavity
¬cavity
catch ¬catch catch ¬catch
0.108 0.012 0.072 0.008
0.016 0.064 0.144 0.576
By Bal Krishna Subedi 17
Artificial Intelligence Reasoning with Uncertainty
Normalization
By Bal Krishna Subedi 18
For distribution:
P(Cavity|toothache) = <P(cavity|toothache), P(¬cavity|toothache)>
= <P(cavity∩toothache)/P(toothache),P(¬cavity∩toothache)/P(toothache)>
= 1/P(toothache) <P(cavity∩ toothache), P(¬cavity∩ toothache)>
= α <P(cavity∩ toothache), P(¬cavity∩ toothache)>
= α P(Cavity, toothache)
= α <0.108+0.012, 0.016+0.064> = α <0.12, 0.08> = <0.6, 0.4>
where α*0.12 + α*0.08 = 1, that is, α = 1/(0.12+0.08) = 5
α can be viewed as a normalization constant for the distribution
P(Cavity|toothache), ensuring that it adds up to 1.
toothache ¬toothache
cavity
¬cavity
catch ¬catch catch ¬catch
0.108 0.012 0.072 0.008
0.016 0.064 0.144 0.576
Artificial Intelligence Reasoning with Uncertainty
General inference procedure
Let X be the query variable (Cavity in the example)
Let E be the set of evidence variables (Toothache in the example),
and e the observed values for them.
Let Y be the remaining unobserved variables (Catch in the example)
The query P(X|e) can be evaluated as follows:
P(X|e) = α P(X, e) = α Σ P(X, e, y)
where the summation is over all possible ys (i.e., all possible
combinations of values of the unobserved variables Y).
Notice that X, E, and Y constitute the complete set of variables for
the domain, so P(X, e, y) is simply a subset of probabilities from the
full joint distribution.
By Bal Krishna Subedi 19
Artificial Intelligence Reasoning with Uncertainty
Independence
A and B are independent iff
P(A|B) = P(A) or P(B|A) = P(B) or P(A, B) = P(A)P(B)
P(Toothache,Catch,Cavity,Weather) = P(Toothache,Catch,Cavity)P(Weather)
By Bal Krishna Subedi 20
Artificial Intelligence Reasoning with Uncertainty
Bayes' Rule
Product rule: Bayes rule:
P(a ^ b) = P(a|b)P(b)
P(a ^ b) = P(b|a)P(a) P(b|a) =
P(a|b) * P(b)
P(a)
By Bal Krishna Subedi 21
Why is the Bayes’ rule is useful in practice ?
Bayes’ rule is useful in practice because there are many cases where
we have good probability estimates for three of the four probabilities
involved, and therefore can compute the fourth one.
Useful for assessing diagnostic probability from causal probability:
P(Cause|Effect) = P(Effect|Cause)P(Cause)
P(Effect)
Diagnostic knowledge is often more fragile than causal knowledge.
Artificial Intelligence Reasoning with Uncertainty
Applying Bayes' Rule
Example:
A doctor knows that the disease meningitis causes the patient to
have a stiff neck 50% of the time. The doctor also knows that the
probability that a patient has meningitis is 1/50,000, and the
probability that any patient has a stiff neck is 1/20.
Find the probability that a patient with a stiff neck has meningitis.
p(s|m) = 0.5
p(m) = 1/50000
p(s) = 1/20
P(m|s) = P(s|m)P(m)/P(s) = (0.5*1/50000)/(1/20) = 0.0002
By Bal Krishna Subedi 22
Artificial Intelligence Reasoning with Uncertainty
Full joint probability distribution: Problem
Full joint probability distribution can answer any question
about the domain, but can become interactively large as the
number of variable grows.
Specifying probability for atomic events is unnatural and
can be very difficult unless large amount of data is available
from which to gather statistical estimates.
By Bal Krishna Subedi 23
Artificial Intelligence Reasoning with Uncertainty
Bayesian networks
Reading: Section 14.1, 14.2, 14.7 from Textbook R&N
A data structure to represent the dependencies among
variables and to give a concise specification of
any full joint probability
distribution.
Also called belief networks or probabilistic network or
casual network or knowledge map.
By Bal Krishna Subedi 24
Artificial Intelligence Reasoning with Uncertainty
Bayesian networks: Definition
A Bayesian network is a directed acyclic graph which consists of:
– A set of random variables which makes up the nodes of the network.
– A set of directed links (arrows) connecting pairs of nodes. If there is an arrow
from node X to node Y, X is said to be a parent of Y.
– Each node Xi has a conditional probability distribution P(Xi | Parents(Xi)) that
quantifies the effect of the parents on the node.
Intuitions:
A Bayesian network models our incomplete understanding of
the causal relationships from an application domain.
A node represents some state of affairs or event.
A link from X to Y means that X has a direct influence on Y.
By Bal Krishna Subedi 25
Artificial Intelligence Reasoning with Uncertainty
Bayesian networks: Example
Our existing simple world of variables toothache, cavity, catch &
weather is represented as:
Weather is independent of the other variables
Toothache and Catch are conditionally independent given Cavity
By Bal Krishna Subedi 26
Artificial Intelligence Reasoning with Uncertainty
Bayesian networks: Example
Sample Domain:
You have a burglar alarm installed in your home. It is fairly reliable at
detecting a burglary, but also responds on occasion to minor
earthquakes. You also have two neighbors, John and Mary, who have
promised to call you at work when they hear the alarm. John always
calls when he hears the alarm, but sometimes confuses the telephone
ringing with the alarm and calls then, too. Mary, on the other hand,
likes rather loud music and sometimes misses the alarm altogether.
We would like have to estimate the probability of a burglary with
given evidence who has or has not call.
Variables: Burglary, Earthquake, Alarm, JohnCalls, MaryCalls
By Bal Krishna Subedi 27
Artificial Intelligence Reasoning with Uncertainty
Bayesian networks: Example
The probabilities associated with the nodes reflect our representation
of the causal relationships.
By Bal Krishna Subedi 28
Artificial Intelligence Reasoning with Uncertainty
Bayesian networks: Semantics
A Bayesian network provides a
complete description of the
domain in the sense that one can
compute the probability of any
state of the world (represented as a
particular assignment to each
variable).
Example: What is the probability that the alarm has sounded, but neither
burglary nor an earthquake has occurred, and both John and Mary call?
By Bal Krishna Subedi 29
P(j, m, a, ¬b, ¬e) = P(j|a) P(m|a) P(a|, ¬b, ¬e) P(¬b) P(¬e) =
= 0.90*0.70*0.001*0.999*0.998 = 0.00062
In general:
P(X1=x1 , … , Xn=xn) = P(x1, … , xn) = Π P(xi|parents(Xi))
i=1
n
Inference in Bayesian networks:
Illustration
Artificial Intelligence Reasoning with Uncertainty
P(b|j, m) = α Σ Σ P(b, e, a, j, m) = α Σ Σ P(b) P(e) P(a|b,e) P(j|a) P(m|a))
= α P(b) ΣP(e) Σ P(a|b,e) P(j|a) P(m|a))
= α*P(b)*ΣP(e)*[P(a|b,e)*P(j|a)*P(m|a) + P(¬a|b,e)*P(j|¬a)*P(m|¬a)] =
= α*P(b)*{P(e)*[P(a|b,e)*P(j|a)*P(m|a) + P(¬a|b,e)*P(j|¬a)*P(m|¬a)] +
P(¬e)*[P(a|b,¬e)*P(j|a)*P(m|a) + P(¬a|b,¬e)*P(j|¬a)*P(m|¬a)]}
= α*0.001*{0.002*[0.95*0.90*0.70 + 0.05*0.05*0.01] +
0.998*[0.94*0.90*0.70 + 0.06*0.05*0.01] }
= α*0.001*{0.002*0.598525 + 0.998*0.59223}
= α*0.001*0.59224259
= α*0.00059224259 = 0.284171835
Example:
Assume the event in which John and Mary called. What is the probability that a
burglary has occurred?
Where α = 1/(0.00059224259 + 0.001491858)
= 479.82337
e a
By Bal Krishna Subedi 30
Artificial Intelligence Reasoning with Uncertainty
Constructing Bayesian networks
1. Choose an ordering of variables X1, … ,Xn
2. For i = 1 to n
– add Xi to the network
– select parents from X1, … ,Xi-1 such that
P (Xi | Parents(Xi)) = P (Xi | X1, ... Xi-1)
This choice of parents guarantees:
P (X1, … ,Xn) = πi =1 P (Xi | X1, … , Xi-1) (chain rule)
= πi =1P (Xi | Parents(Xi)) (by construction)
By Bal Krishna Subedi 31
Constructing Bayesian networks:
Example
Artificial Intelligence Reasoning with Uncertainty
Suppose we choose the ordering MarryCall, JohnCall, Alarm,
Burglary, Earthquake.
• Adding MarryCalls: No
parents
• Adding JohnCalls: if Marry
calls, probably alarm is gone
off, which likely causes
John calls. Therefore, John
needs Marry call as parent.
By Bal Krishna Subedi 32
Constructing Bayesian networks:
Example
Artificial Intelligence Reasoning with Uncertainty
• Adding Alarm: If both calls it is
more likely that the alarm has
gone off, so we need both
JohnCall and MerryCall as
parents.
By Bal Krishna Subedi 33
Constructing Bayesian networks:
Example
Artificial Intelligence Reasoning with Uncertainty
• Adding Burglary: If the
alarm is on there has been a
burglary. In alarm state
MarryCalls, JohnCalls gives
information about phone
ringing ot Marry’s music,
but not about the burglary.
Hence just as parent.
By Bal Krishna Subedi 34
Constructing Bayesian networks:
Example
Artificial Intelligence Reasoning with Uncertainty
• Adding Earthquake: If
the alarm is on, it is
more likely that there
has been an earthquake.
But if there is burglary,
then that explains the
alarm, and the
probability would be
above the normal.
By Bal Krishna Subedi 35
Resulting network has two more links than the original and requires three more
probabilities to be specified .
Node ordering is important issue in Bayesian Network construction.
Other Approaches to Uncertain
Reasoning
Artificial Intelligence Reasoning with Uncertainty
Default or nonmonotonic logic:
“believed until a better reason found to believe something else”
Assume my car does not have a flat tire
Assume A25 works unless contradicted by evidence
Issues: What assumptions are reasonable? How to handle contradiction?
Rules with fudge factors:
A25 0.3 get there on time
Sprinkler 0.99 WetGrass
WetGrass 0.7 Rain
Issues: Problems with combination, e.g., Sprinkler causes Rain??
Fuzzy Logic:
Fuzzy logic handles degree of truth NOT uncertainty e.g.,
WetGrass is true to degree 0.2)
By Bal Krishna Subedi 36
Artificial Intelligence Reasoning with Uncertainty
Fuzzy Logic
Fuzzy logic is a method for reasoning with logical expressions
describing membership in fuzzy sets.
Fuzzy predicate and says that the truth value in a number between 0 & 1.
Example: “John is tall” is true if his height is 78?
We can express as Tall(John) = 0.6 true
Handaling complex sentences:
T(A^B) = min(T(A), T(B))
T(AVB) = max(T(A),T(B))
T(¬A) = 1- T(A)
Suppose Tall(John) = 0.6 and Heavy(John) = 0.4
T(Tall(John)) ^ T(Heavy(John)) = 0.4
By Bal Krishna Subedi 37
Artificial Intelligence Reasoning with Uncertainty
Home Works
1. Problems 13.1, 13.2, 13.5, 13.6 & 13.8
2. Problems 14.1 & 14.2
3. Create a Bayesian network for application
automobile fault analysis.
By Bal Krishna Subedi 38

More Related Content

What's hot

Problem solving agents
Problem solving agentsProblem solving agents
Problem solving agentsMegha Sharma
 
Operating system Dead lock
Operating system Dead lockOperating system Dead lock
Operating system Dead lockKaram Munir Butt
 
INTRODUCTION TO ALGORITHMS Third Edition
INTRODUCTION TO ALGORITHMS Third EditionINTRODUCTION TO ALGORITHMS Third Edition
INTRODUCTION TO ALGORITHMS Third EditionPHI Learning Pvt. Ltd.
 
Informed search (heuristics)
Informed search (heuristics)Informed search (heuristics)
Informed search (heuristics)Bablu Shofi
 
Dynamic programming, Branch and bound algorithm & Greedy algorithms
Dynamic programming, Branch and bound algorithm & Greedy algorithms Dynamic programming, Branch and bound algorithm & Greedy algorithms
Dynamic programming, Branch and bound algorithm & Greedy algorithms SURBHI SAROHA
 
Production System in AI
Production System in AIProduction System in AI
Production System in AIBharat Bhushan
 
State space search
State space searchState space search
State space searchchauhankapil
 
I. AO* SEARCH ALGORITHM
I. AO* SEARCH ALGORITHMI. AO* SEARCH ALGORITHM
I. AO* SEARCH ALGORITHMvikas dhakane
 
Control Strategies in AI
Control Strategies in AI Control Strategies in AI
Control Strategies in AI Bharat Bhushan
 
Unification and Lifting
Unification and LiftingUnification and Lifting
Unification and LiftingMegha Sharma
 
data structures- back tracking
data structures- back trackingdata structures- back tracking
data structures- back trackingAbinaya B
 
Knowledge Representation in Artificial intelligence
Knowledge Representation in Artificial intelligence Knowledge Representation in Artificial intelligence
Knowledge Representation in Artificial intelligence Yasir Khan
 
Hill climbing algorithm in artificial intelligence
Hill climbing algorithm in artificial intelligenceHill climbing algorithm in artificial intelligence
Hill climbing algorithm in artificial intelligencesandeep54552
 

What's hot (20)

Problem solving agents
Problem solving agentsProblem solving agents
Problem solving agents
 
Operating system Dead lock
Operating system Dead lockOperating system Dead lock
Operating system Dead lock
 
AI Lecture 7 (uncertainty)
AI Lecture 7 (uncertainty)AI Lecture 7 (uncertainty)
AI Lecture 7 (uncertainty)
 
INTRODUCTION TO ALGORITHMS Third Edition
INTRODUCTION TO ALGORITHMS Third EditionINTRODUCTION TO ALGORITHMS Third Edition
INTRODUCTION TO ALGORITHMS Third Edition
 
Planning
PlanningPlanning
Planning
 
AI: Logic in AI
AI: Logic in AIAI: Logic in AI
AI: Logic in AI
 
Problems, Problem spaces and Search
Problems, Problem spaces and SearchProblems, Problem spaces and Search
Problems, Problem spaces and Search
 
Informed search (heuristics)
Informed search (heuristics)Informed search (heuristics)
Informed search (heuristics)
 
Dynamic programming, Branch and bound algorithm & Greedy algorithms
Dynamic programming, Branch and bound algorithm & Greedy algorithms Dynamic programming, Branch and bound algorithm & Greedy algorithms
Dynamic programming, Branch and bound algorithm & Greedy algorithms
 
Production System in AI
Production System in AIProduction System in AI
Production System in AI
 
State space search
State space searchState space search
State space search
 
I. AO* SEARCH ALGORITHM
I. AO* SEARCH ALGORITHMI. AO* SEARCH ALGORITHM
I. AO* SEARCH ALGORITHM
 
Control Strategies in AI
Control Strategies in AI Control Strategies in AI
Control Strategies in AI
 
Knowledge based agents
Knowledge based agentsKnowledge based agents
Knowledge based agents
 
Hill climbing algorithm
Hill climbing algorithmHill climbing algorithm
Hill climbing algorithm
 
Unification and Lifting
Unification and LiftingUnification and Lifting
Unification and Lifting
 
data structures- back tracking
data structures- back trackingdata structures- back tracking
data structures- back tracking
 
5 csp
5 csp5 csp
5 csp
 
Knowledge Representation in Artificial intelligence
Knowledge Representation in Artificial intelligence Knowledge Representation in Artificial intelligence
Knowledge Representation in Artificial intelligence
 
Hill climbing algorithm in artificial intelligence
Hill climbing algorithm in artificial intelligenceHill climbing algorithm in artificial intelligence
Hill climbing algorithm in artificial intelligence
 

Similar to Uncertainty

Uncertain knowledge and reasoning
Uncertain knowledge and reasoningUncertain knowledge and reasoning
Uncertain knowledge and reasoningShiwani Gupta
 
Probability Theory.pdf
Probability Theory.pdfProbability Theory.pdf
Probability Theory.pdfcodewithmensah
 
Bayes Classification
Bayes ClassificationBayes Classification
Bayes Classificationsathish sak
 
Basic Concept Of Probability
Basic Concept Of ProbabilityBasic Concept Of Probability
Basic Concept Of Probabilityguest45a926
 
Probability concepts for Data Analytics
Probability concepts for Data AnalyticsProbability concepts for Data Analytics
Probability concepts for Data AnalyticsSSaudia
 
Equational axioms for probability calculus and modelling of Likelihood ratio ...
Equational axioms for probability calculus and modelling of Likelihood ratio ...Equational axioms for probability calculus and modelling of Likelihood ratio ...
Equational axioms for probability calculus and modelling of Likelihood ratio ...Advanced-Concepts-Team
 
MATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptx
MATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptxMATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptx
MATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptxpavantech57
 
Jarrar.lecture notes.aai.2011s.ch7.p logic
Jarrar.lecture notes.aai.2011s.ch7.p logicJarrar.lecture notes.aai.2011s.ch7.p logic
Jarrar.lecture notes.aai.2011s.ch7.p logicPalGov
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheetSuvrat Mishra
 
Probability Review Additions
Probability Review AdditionsProbability Review Additions
Probability Review Additionsrishi.indian
 
AI 8 | Probability Basics, Bayes' Rule, Probability Distribution
AI 8 | Probability Basics, Bayes' Rule, Probability DistributionAI 8 | Probability Basics, Bayes' Rule, Probability Distribution
AI 8 | Probability Basics, Bayes' Rule, Probability DistributionMohammad Imam Hossain
 
Data mining assignment 2
Data mining assignment 2Data mining assignment 2
Data mining assignment 2BarryK88
 

Similar to Uncertainty (20)

Uncertain knowledge and reasoning
Uncertain knowledge and reasoningUncertain knowledge and reasoning
Uncertain knowledge and reasoning
 
AI_7 Statistical Reasoning
AI_7 Statistical Reasoning AI_7 Statistical Reasoning
AI_7 Statistical Reasoning
 
Uncertainity
Uncertainity Uncertainity
Uncertainity
 
Probability Theory.pdf
Probability Theory.pdfProbability Theory.pdf
Probability Theory.pdf
 
Chapter 13
Chapter 13Chapter 13
Chapter 13
 
Bayes Classification
Bayes ClassificationBayes Classification
Bayes Classification
 
HPWFcorePRES--FUR2016
HPWFcorePRES--FUR2016HPWFcorePRES--FUR2016
HPWFcorePRES--FUR2016
 
Basic Concept Of Probability
Basic Concept Of ProbabilityBasic Concept Of Probability
Basic Concept Of Probability
 
Probability concepts for Data Analytics
Probability concepts for Data AnalyticsProbability concepts for Data Analytics
Probability concepts for Data Analytics
 
Bayesian statistics
Bayesian statisticsBayesian statistics
Bayesian statistics
 
Equational axioms for probability calculus and modelling of Likelihood ratio ...
Equational axioms for probability calculus and modelling of Likelihood ratio ...Equational axioms for probability calculus and modelling of Likelihood ratio ...
Equational axioms for probability calculus and modelling of Likelihood ratio ...
 
PTSP PPT.pdf
PTSP PPT.pdfPTSP PPT.pdf
PTSP PPT.pdf
 
MATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptx
MATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptxMATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptx
MATHS PRESENTATION OF STATISTICS AND PROBABILITY.pptx
 
Jarrar.lecture notes.aai.2011s.ch7.p logic
Jarrar.lecture notes.aai.2011s.ch7.p logicJarrar.lecture notes.aai.2011s.ch7.p logic
Jarrar.lecture notes.aai.2011s.ch7.p logic
 
Probability cheatsheet
Probability cheatsheetProbability cheatsheet
Probability cheatsheet
 
Probability Cheatsheet.pdf
Probability Cheatsheet.pdfProbability Cheatsheet.pdf
Probability Cheatsheet.pdf
 
Probability Review Additions
Probability Review AdditionsProbability Review Additions
Probability Review Additions
 
13-uncertainty.pdf
13-uncertainty.pdf13-uncertainty.pdf
13-uncertainty.pdf
 
AI 8 | Probability Basics, Bayes' Rule, Probability Distribution
AI 8 | Probability Basics, Bayes' Rule, Probability DistributionAI 8 | Probability Basics, Bayes' Rule, Probability Distribution
AI 8 | Probability Basics, Bayes' Rule, Probability Distribution
 
Data mining assignment 2
Data mining assignment 2Data mining assignment 2
Data mining assignment 2
 

More from Tribhuvan University (20)

Lecture 7 se
Lecture 7 seLecture 7 se
Lecture 7 se
 
Lecture 6 se
Lecture 6 seLecture 6 se
Lecture 6 se
 
Lecture 5 se
Lecture 5 seLecture 5 se
Lecture 5 se
 
Lecture 3 se
Lecture 3 seLecture 3 se
Lecture 3 se
 
Lecture 1 se
Lecture 1 seLecture 1 se
Lecture 1 se
 
Lecture 2 se
Lecture 2 seLecture 2 se
Lecture 2 se
 
Lecture 5 m&amp;ca
Lecture 5 m&amp;caLecture 5 m&amp;ca
Lecture 5 m&amp;ca
 
Lecture 4 m&amp;ca
Lecture 4 m&amp;caLecture 4 m&amp;ca
Lecture 4 m&amp;ca
 
Lecture 3 m&amp;ca
Lecture 3 m&amp;caLecture 3 m&amp;ca
Lecture 3 m&amp;ca
 
Lecture 2 m&amp;ca
Lecture 2 m&amp;caLecture 2 m&amp;ca
Lecture 2 m&amp;ca
 
Lecture 1 m&amp;ca
Lecture 1 m&amp;caLecture 1 m&amp;ca
Lecture 1 m&amp;ca
 
Neural network (csc372) lecture 2
Neural network (csc372) lecture 2Neural network (csc372) lecture 2
Neural network (csc372) lecture 2
 
Neural network (csc372) lecture 1
Neural network (csc372) lecture 1Neural network (csc372) lecture 1
Neural network (csc372) lecture 1
 
Ai unit-2
Ai unit-2Ai unit-2
Ai unit-2
 
Ai unit-3
Ai unit-3Ai unit-3
Ai unit-3
 
Ai unit-4
Ai unit-4Ai unit-4
Ai unit-4
 
Unit 4(nlp _neural_network)
Unit 4(nlp _neural_network)Unit 4(nlp _neural_network)
Unit 4(nlp _neural_network)
 
Ai unit-6
Ai unit-6Ai unit-6
Ai unit-6
 
Ai unit-1
Ai unit-1Ai unit-1
Ai unit-1
 
Knowldge reprsentations
Knowldge reprsentationsKnowldge reprsentations
Knowldge reprsentations
 

Recently uploaded

General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024Janet Corral
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDThiyagu K
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingTeacherCyreneCayanan
 
Student login on Anyboli platform.helpin
Student login on Anyboli platform.helpinStudent login on Anyboli platform.helpin
Student login on Anyboli platform.helpinRaunakKeshri1
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Disha Kariya
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfciinovamais
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxVishalSingh1417
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...fonyou31
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104misteraugie
 
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...PsychoTech Services
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformChameera Dedduwage
 

Recently uploaded (20)

General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writing
 
Student login on Anyboli platform.helpin
Student login on Anyboli platform.helpinStudent login on Anyboli platform.helpin
Student login on Anyboli platform.helpin
 
Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..Sports & Fitness Value Added Course FY..
Sports & Fitness Value Added Course FY..
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104Nutritional Needs Presentation - HLTH 104
Nutritional Needs Presentation - HLTH 104
 
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
IGNOU MSCCFT and PGDCFT Exam Question Pattern: MCFT003 Counselling and Family...
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
A Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy ReformA Critique of the Proposed National Education Policy Reform
A Critique of the Proposed National Education Policy Reform
 

Uncertainty

  • 1. Artificial Intelligence Reasoning with Uncertainty Uncertainty Reading: Section 13.1 - 13.6 of Textbook R&N In which we see what an agent should do when not all is crystal clear. By Bal Krishna Subedi 1
  • 2. Artificial Intelligence Reasoning with Uncertainty Uncertain Knowledge Let action At = leave for airport t minutes before flight Will At get me there on time? Problems: 1) partial observability (road state, other drivers' plans, etc.) 2) noisy sensors (radio traffic reports) 3) uncertainty in action outcomes (flat tyre, etc.) 4) complexity of modeling and predicting traffic Hence a purely logical approach either 1) risks falsehood: “A25 will get me there on time” or 2) leads to conclusions that are too weak for decision making: “A25 will get me there on time if there's no accident on the bridge and it doesn't rain and my tires remain intact etc etc.” A1440 might reasonably be said to get me there on time but I'd have to stay overnight in the airport… By Bal Krishna Subedi 2
  • 3. Artificial Intelligence Reasoning with Uncertainty Handling Uncertainty Instead of providing all condition it can express with degree of beliefs in the relevant sentences. Example: Say we have a rule if toothache then problem is cavity But not all patients have toothaches because of cavities (although perhaps most do) So we could set up rules like if toothache and not(gum disease) and not(filling) and ......then problem is cavity This gets very complicated! a better method would be to say if toothache then problem is cavity with probability 0.8 Given the available evidence, A25 will get me there on time with probability 0.04 A most important tool for dealing with degree of beliefs is probability theory, which assigns to each sentence a numerical degree of belief between 0 & 1. By Bal Krishna Subedi 3
  • 4. Artificial Intelligence Reasoning with Uncertainty Making decisions under uncertainty Suppose I believe the following: P(A25 gets me there on time|…) = 0.04 P(A90 gets me there on time|…) = 0.70 P(A120 gets me there on time|…) = 0.95 P(A1440 gets me there on time|…) = 0.9999 Which action to choose? Depends on my preferences for missing flight vs. length of wait at airport, etc. Utility theory is used to represent and infer preferences Decision theory = utility theory + probability theory The rational decision depends on both the relative importance of various goals and the likelihood that, and degree to which, they will be achieved. By Bal Krishna Subedi 4
  • 5. Artificial Intelligence Reasoning with Uncertainty Probability basics Begin with a set Ω - the sample space e.g., 6 possible rolls of a die. ω ε Ω is a sample point/possible world/atomic event A probability space or probability model is a sample space with an assignment P(ω) for every ω ε Ω s.t. 0 ≤ P(ω) ≥ 1 ∑P(ω) = 1 e.g., P(1) = P(2) = P(3) = P(4) = P(5) = P(6) = 1/6. An event A is any subset of Ω P(A) = ∑{ωεA} P(ω) E.g., P(die roll < 4) = 1/6 + 1/6 + 1/6 = 1/2 By Bal Krishna Subedi 5
  • 6. Artificial Intelligence Reasoning with Uncertainty Random variables By Bal Krishna Subedi 6 An application domain (a world) is described by a set of random variables, each variable referring to a part of the domain whose status is initially unknown. Example: A simple world consisting of two random variables: Cavity – a Boolean variable that refers to whether my lower left wisdom tooth has a cavity Toothache - a Boolean variable that refers to whether I have a toothache or not. We use the single capital letters to represent unknown random variables P induces a probability distribution for any r.v. X: P(X = xi) = ∑ {ω: X(ω) = xi}P(ω) e.g., P(Odd = true) = 1/6 + 1/6 + 1/6 = 1/2 Each RV has a domain of values that it can take it. e. g. domain of Cavity is {true, false} RVs domain are: Boolean, Discrete and Continiuous
  • 7. Artificial Intelligence Reasoning with Uncertainty Atomic events An atomic event is a complete specification of the state of the world about which the agent is uncertain. Example: In the above world with two random variables (Cavity and Toothache) there are only four distinct atomic events, one being: Cavity = false, Toothache = true Which are the other three atomic events? By Bal Krishna Subedi 7
  • 8. Artificial Intelligence Reasoning with Uncertainty Propositions Think of a proposition as the event (set of sample points) where the proposition is true Given Boolean random variables A and B: event α = set of sample points where A(ω) = true event ¬α = set of sample points where A(ω) = false event a ^ b = points where A(ω) = true and B(ω) = true Often in AI applications, the sample points are defined by the values of a set of random variables, i.e., the sample space is the Cartesian product of the ranges of the variables. With Boolean variables, sample point = propositional logic model e.g., A = true, B = false, or a ^ ¬b. Proposition = disjunction of atomic events in which it is true e.g., (a ∨ b) ≡ (¬a ^ b) ∨ (a ^ ¬b) ∨ (a ^ b) P(a ∨ b) = P(¬a ^ b) + P(a ^ ¬b) + P(a ^ b) By Bal Krishna Subedi 8
  • 9. Artificial Intelligence Reasoning with Uncertainty Syntax for propositions Propositional or Boolean random variables e.g., Cavity (do I have a cavity?) Discrete random variables (finite or infinite) e.g., Weather is one of (sunny, rain, cloudy, snow) Weather = rain is a proposition Values must be exhaustive and mutually exclusive Continuous random variables (bounded or unbounded) e.g., Temp = 21.6, also allow, e.g., Temp < 22.0. By Bal Krishna Subedi 9
  • 10. Artificial Intelligence Reasoning with Uncertainty Prior probability The prior or unconditional probability associated with a proposition is the degree of belief accorded to it in the absence of any other information. Example: P(Weather = sunny) = 0.72 P(Cavity = true) = 0.1 or P(cavity) = 0.1 Probability distribution gives values for all possible assignments: P(Weather) = (0.72, 0.1, 0.08, 0.1) Joint probability distribution for a set of r.v.s gives the probability of every atomic event on those r.v.s (i.e., every sample point) P(Weather, Cavity) = a 4 × 2 matrix of values. By Bal Krishna Subedi 10 Every question about a domain can be answered by the joint distribution because every event is a sum of sample points
  • 11. Artificial Intelligence Reasoning with Uncertainty Conditional probability The conditional probability “P(a|b)” is the probability of “a” given that all we know is “b”. e.g., P(cavity|toothache) = 0.8 means if a patient have toothache and no other information is yet available, then the probability of patient’s having the cavity is 0.8 Definition of conditional probability: P(a|b) = P(a ^ b)/P(b) if P(b) ≠ 0 Product rule gives an alternative formulation: P(a ^ b) = P(a|b)P(b) = p(b|a)P(a) Chain rule is derived by successive application of product rule: P(X1,…,Xn) = P(X1,…,Xn-1) P(Xn|X1,…,Xn-1) …….. = ∏n i=1P(Xi|X1,…,Xi-1) By Bal Krishna Subedi 11
  • 12. Artificial Intelligence Reasoning with Uncertainty The axioms of probability 1. All probabilities are between 0 and 1: 0 ≤ P(a) ≤ 1 2. Necessarily true propositions have probability 1 and necessarily false propositions have probability 0: P(true) = 1 P(false) = 0 3. The probability of a disjunction is given by: P(a U b) = P(a) + P(b) – P(a ∩ b) By Bal Krishna Subedi 12
  • 13. Inference using full joint probability distribution Artificial Intelligence Reasoning with Uncertainty We use the full joint distribution as the knowledge base from which answers to all questions may be derived. The probability of a proposition is equal to the sum of the probabilities of the atomic events in which it holds. P(a) = Σ P(ei) Therefore, given a full joint distribution that specifies the probabilities of all the atomic events, one can compute the probability of any proposition. By Bal Krishna Subedi 13
  • 14. Artificial Intelligence Reasoning with Uncertainty Full joint probability distribution: example We consider the following domain consisting of three Boolean variables: Toothache, Cavity, and Catch (the dentist’s nasty steel probe catches in my tooth). The full joint distribution is the following 2x2x2 table: toothache ¬toothache cavity ¬cavity catch ¬catch catch ¬catch 0.108 0.012 0.072 0.008 0.016 0.064 0.144 0.576 The probabilities in the joint distribution must sum to 1. The probability of any proposition can be computed from the probabilities in the table. By Bal Krishna Subedi 14 Each cell represents an atomic event and these are all the possible atomic events.
  • 15. Artificial Intelligence Reasoning with Uncertainty Full joint probability distribution: example toothache ¬toothache cavity ¬cavity catch ¬catch catch ¬catch 0.108 0.012 0.072 0.008 0.016 0.064 0.144 0.576 We simply identify those atomic events in which the proposition is true and add up their probabilities P(cavity or toothache) = P(cavity, toothache, catch) + P(cavity, toothache, ¬catch) + P(cavity, ¬toothache, catch) + P(cavity, ¬toothache, ¬catch) + P(¬cavity, toothache, catch) + P(¬cavity, toothache, ¬catch) = 0.108+0.012+0.072+0.008+0.016+0.064=0.28 By Bal Krishna Subedi 15
  • 16. Artificial Intelligence Reasoning with Uncertainty Marginalization or summing out toothache ¬toothache cavity ¬cavity catch ¬catch catch ¬catch 0.108 0.012 0.072 0.008 0.016 0.064 0.144 0.576 P(cavity) P(¬toothache) P(cavity, ¬toothache) That is, a distribution over Y can be obtained by summing out all the other variables from any joint distribution containing Y. P(Y) = Σ P(Y, z) P(cavity) = 0.108+0.012+0.072+0.008 = 0.2 P(¬toothache) = 0.072+0.008+0.144+0.576 = 0.8 P(cavity, ¬toothache) = 0.072+0.008 = 0.08 By Bal Krishna Subedi 16
  • 17. Artificial Intelligence Reasoning with Uncertainty Conditioning P(Y) = Σ P(Y, z) P(Y, z) = P(Y|z)P(z) Therefore, for any set of variables Y and Z: P(Y) = Σ P(Y|z)P(z) - This is the conditioning rule toothache ¬toothache cavity ¬cavity catch ¬catch catch ¬catch 0.108 0.012 0.072 0.008 0.016 0.064 0.144 0.576 By Bal Krishna Subedi 17
  • 18. Artificial Intelligence Reasoning with Uncertainty Normalization By Bal Krishna Subedi 18 For distribution: P(Cavity|toothache) = <P(cavity|toothache), P(¬cavity|toothache)> = <P(cavity∩toothache)/P(toothache),P(¬cavity∩toothache)/P(toothache)> = 1/P(toothache) <P(cavity∩ toothache), P(¬cavity∩ toothache)> = α <P(cavity∩ toothache), P(¬cavity∩ toothache)> = α P(Cavity, toothache) = α <0.108+0.012, 0.016+0.064> = α <0.12, 0.08> = <0.6, 0.4> where α*0.12 + α*0.08 = 1, that is, α = 1/(0.12+0.08) = 5 α can be viewed as a normalization constant for the distribution P(Cavity|toothache), ensuring that it adds up to 1. toothache ¬toothache cavity ¬cavity catch ¬catch catch ¬catch 0.108 0.012 0.072 0.008 0.016 0.064 0.144 0.576
  • 19. Artificial Intelligence Reasoning with Uncertainty General inference procedure Let X be the query variable (Cavity in the example) Let E be the set of evidence variables (Toothache in the example), and e the observed values for them. Let Y be the remaining unobserved variables (Catch in the example) The query P(X|e) can be evaluated as follows: P(X|e) = α P(X, e) = α Σ P(X, e, y) where the summation is over all possible ys (i.e., all possible combinations of values of the unobserved variables Y). Notice that X, E, and Y constitute the complete set of variables for the domain, so P(X, e, y) is simply a subset of probabilities from the full joint distribution. By Bal Krishna Subedi 19
  • 20. Artificial Intelligence Reasoning with Uncertainty Independence A and B are independent iff P(A|B) = P(A) or P(B|A) = P(B) or P(A, B) = P(A)P(B) P(Toothache,Catch,Cavity,Weather) = P(Toothache,Catch,Cavity)P(Weather) By Bal Krishna Subedi 20
  • 21. Artificial Intelligence Reasoning with Uncertainty Bayes' Rule Product rule: Bayes rule: P(a ^ b) = P(a|b)P(b) P(a ^ b) = P(b|a)P(a) P(b|a) = P(a|b) * P(b) P(a) By Bal Krishna Subedi 21 Why is the Bayes’ rule is useful in practice ? Bayes’ rule is useful in practice because there are many cases where we have good probability estimates for three of the four probabilities involved, and therefore can compute the fourth one. Useful for assessing diagnostic probability from causal probability: P(Cause|Effect) = P(Effect|Cause)P(Cause) P(Effect) Diagnostic knowledge is often more fragile than causal knowledge.
  • 22. Artificial Intelligence Reasoning with Uncertainty Applying Bayes' Rule Example: A doctor knows that the disease meningitis causes the patient to have a stiff neck 50% of the time. The doctor also knows that the probability that a patient has meningitis is 1/50,000, and the probability that any patient has a stiff neck is 1/20. Find the probability that a patient with a stiff neck has meningitis. p(s|m) = 0.5 p(m) = 1/50000 p(s) = 1/20 P(m|s) = P(s|m)P(m)/P(s) = (0.5*1/50000)/(1/20) = 0.0002 By Bal Krishna Subedi 22
  • 23. Artificial Intelligence Reasoning with Uncertainty Full joint probability distribution: Problem Full joint probability distribution can answer any question about the domain, but can become interactively large as the number of variable grows. Specifying probability for atomic events is unnatural and can be very difficult unless large amount of data is available from which to gather statistical estimates. By Bal Krishna Subedi 23
  • 24. Artificial Intelligence Reasoning with Uncertainty Bayesian networks Reading: Section 14.1, 14.2, 14.7 from Textbook R&N A data structure to represent the dependencies among variables and to give a concise specification of any full joint probability distribution. Also called belief networks or probabilistic network or casual network or knowledge map. By Bal Krishna Subedi 24
  • 25. Artificial Intelligence Reasoning with Uncertainty Bayesian networks: Definition A Bayesian network is a directed acyclic graph which consists of: – A set of random variables which makes up the nodes of the network. – A set of directed links (arrows) connecting pairs of nodes. If there is an arrow from node X to node Y, X is said to be a parent of Y. – Each node Xi has a conditional probability distribution P(Xi | Parents(Xi)) that quantifies the effect of the parents on the node. Intuitions: A Bayesian network models our incomplete understanding of the causal relationships from an application domain. A node represents some state of affairs or event. A link from X to Y means that X has a direct influence on Y. By Bal Krishna Subedi 25
  • 26. Artificial Intelligence Reasoning with Uncertainty Bayesian networks: Example Our existing simple world of variables toothache, cavity, catch & weather is represented as: Weather is independent of the other variables Toothache and Catch are conditionally independent given Cavity By Bal Krishna Subedi 26
  • 27. Artificial Intelligence Reasoning with Uncertainty Bayesian networks: Example Sample Domain: You have a burglar alarm installed in your home. It is fairly reliable at detecting a burglary, but also responds on occasion to minor earthquakes. You also have two neighbors, John and Mary, who have promised to call you at work when they hear the alarm. John always calls when he hears the alarm, but sometimes confuses the telephone ringing with the alarm and calls then, too. Mary, on the other hand, likes rather loud music and sometimes misses the alarm altogether. We would like have to estimate the probability of a burglary with given evidence who has or has not call. Variables: Burglary, Earthquake, Alarm, JohnCalls, MaryCalls By Bal Krishna Subedi 27
  • 28. Artificial Intelligence Reasoning with Uncertainty Bayesian networks: Example The probabilities associated with the nodes reflect our representation of the causal relationships. By Bal Krishna Subedi 28
  • 29. Artificial Intelligence Reasoning with Uncertainty Bayesian networks: Semantics A Bayesian network provides a complete description of the domain in the sense that one can compute the probability of any state of the world (represented as a particular assignment to each variable). Example: What is the probability that the alarm has sounded, but neither burglary nor an earthquake has occurred, and both John and Mary call? By Bal Krishna Subedi 29 P(j, m, a, ¬b, ¬e) = P(j|a) P(m|a) P(a|, ¬b, ¬e) P(¬b) P(¬e) = = 0.90*0.70*0.001*0.999*0.998 = 0.00062 In general: P(X1=x1 , … , Xn=xn) = P(x1, … , xn) = Π P(xi|parents(Xi)) i=1 n
  • 30. Inference in Bayesian networks: Illustration Artificial Intelligence Reasoning with Uncertainty P(b|j, m) = α Σ Σ P(b, e, a, j, m) = α Σ Σ P(b) P(e) P(a|b,e) P(j|a) P(m|a)) = α P(b) ΣP(e) Σ P(a|b,e) P(j|a) P(m|a)) = α*P(b)*ΣP(e)*[P(a|b,e)*P(j|a)*P(m|a) + P(¬a|b,e)*P(j|¬a)*P(m|¬a)] = = α*P(b)*{P(e)*[P(a|b,e)*P(j|a)*P(m|a) + P(¬a|b,e)*P(j|¬a)*P(m|¬a)] + P(¬e)*[P(a|b,¬e)*P(j|a)*P(m|a) + P(¬a|b,¬e)*P(j|¬a)*P(m|¬a)]} = α*0.001*{0.002*[0.95*0.90*0.70 + 0.05*0.05*0.01] + 0.998*[0.94*0.90*0.70 + 0.06*0.05*0.01] } = α*0.001*{0.002*0.598525 + 0.998*0.59223} = α*0.001*0.59224259 = α*0.00059224259 = 0.284171835 Example: Assume the event in which John and Mary called. What is the probability that a burglary has occurred? Where α = 1/(0.00059224259 + 0.001491858) = 479.82337 e a By Bal Krishna Subedi 30
  • 31. Artificial Intelligence Reasoning with Uncertainty Constructing Bayesian networks 1. Choose an ordering of variables X1, … ,Xn 2. For i = 1 to n – add Xi to the network – select parents from X1, … ,Xi-1 such that P (Xi | Parents(Xi)) = P (Xi | X1, ... Xi-1) This choice of parents guarantees: P (X1, … ,Xn) = πi =1 P (Xi | X1, … , Xi-1) (chain rule) = πi =1P (Xi | Parents(Xi)) (by construction) By Bal Krishna Subedi 31
  • 32. Constructing Bayesian networks: Example Artificial Intelligence Reasoning with Uncertainty Suppose we choose the ordering MarryCall, JohnCall, Alarm, Burglary, Earthquake. • Adding MarryCalls: No parents • Adding JohnCalls: if Marry calls, probably alarm is gone off, which likely causes John calls. Therefore, John needs Marry call as parent. By Bal Krishna Subedi 32
  • 33. Constructing Bayesian networks: Example Artificial Intelligence Reasoning with Uncertainty • Adding Alarm: If both calls it is more likely that the alarm has gone off, so we need both JohnCall and MerryCall as parents. By Bal Krishna Subedi 33
  • 34. Constructing Bayesian networks: Example Artificial Intelligence Reasoning with Uncertainty • Adding Burglary: If the alarm is on there has been a burglary. In alarm state MarryCalls, JohnCalls gives information about phone ringing ot Marry’s music, but not about the burglary. Hence just as parent. By Bal Krishna Subedi 34
  • 35. Constructing Bayesian networks: Example Artificial Intelligence Reasoning with Uncertainty • Adding Earthquake: If the alarm is on, it is more likely that there has been an earthquake. But if there is burglary, then that explains the alarm, and the probability would be above the normal. By Bal Krishna Subedi 35 Resulting network has two more links than the original and requires three more probabilities to be specified . Node ordering is important issue in Bayesian Network construction.
  • 36. Other Approaches to Uncertain Reasoning Artificial Intelligence Reasoning with Uncertainty Default or nonmonotonic logic: “believed until a better reason found to believe something else” Assume my car does not have a flat tire Assume A25 works unless contradicted by evidence Issues: What assumptions are reasonable? How to handle contradiction? Rules with fudge factors: A25 0.3 get there on time Sprinkler 0.99 WetGrass WetGrass 0.7 Rain Issues: Problems with combination, e.g., Sprinkler causes Rain?? Fuzzy Logic: Fuzzy logic handles degree of truth NOT uncertainty e.g., WetGrass is true to degree 0.2) By Bal Krishna Subedi 36
  • 37. Artificial Intelligence Reasoning with Uncertainty Fuzzy Logic Fuzzy logic is a method for reasoning with logical expressions describing membership in fuzzy sets. Fuzzy predicate and says that the truth value in a number between 0 & 1. Example: “John is tall” is true if his height is 78? We can express as Tall(John) = 0.6 true Handaling complex sentences: T(A^B) = min(T(A), T(B)) T(AVB) = max(T(A),T(B)) T(¬A) = 1- T(A) Suppose Tall(John) = 0.6 and Heavy(John) = 0.4 T(Tall(John)) ^ T(Heavy(John)) = 0.4 By Bal Krishna Subedi 37
  • 38. Artificial Intelligence Reasoning with Uncertainty Home Works 1. Problems 13.1, 13.2, 13.5, 13.6 & 13.8 2. Problems 14.1 & 14.2 3. Create a Bayesian network for application automobile fault analysis. By Bal Krishna Subedi 38