SlideShare a Scribd company logo
CSC411 Artificial Intelligence 1
Chapter 9
Reasoning in Uncertain Situations
Contents
Uncertain situations
Non-monotonic logic and reasoning
Certainty Factor algebra
Fuzzy logic and reasoning
Dempster-Shafer theory of evidence
Bayesian belief network
Markov models
CSC411 Artificial Intelligence 2
Traditional Logic
Based on predicate logic
Three important assumptions:
– Predicate descriptions are sufficient
w.r.t. to the domain
– Information is consistent
– Knowledge base grows monotonically
CSC411 Artificial Intelligence 3
Non-monotonic Logic
Addresses the three assumptions of
traditional logic
– Knowledge is incomplete
No knowledge about p: true or false?
Prolog – closed world assumption
– Knowledge is inconsistent
Based on how the world usually works
Most birds fly, but Ostrich doesn’t
– Knowledge base grows non-monotonically
New observation may contradict the existing
knowledge, thus the existing knowledge may need
removal.
Inference based on assumptions, how come if the
assumptions are later shown to be incorrect
Three modal operators are introduced
CSC411 Artificial Intelligence 4
Unless Operator
New information may invalidate previous results
Implemented in TMS – Truth Maintenance
Systems to keep track of the reasoning steps and
preserve the KB consistency
Introduce Unless operator
– Support inferences based on the belief that its argument
is not true
– Consider
p(X) unless q(X)  r(X)
If p(X) is true and not believe q(X) true then r(X)
p(Z)
r(W)  s(W)
From above, conclude s(X).
Later, change believe or find q(X) true, what happens?
Retract r(X) and s(X)
– Unless deals with believe, not truth
Either unknown or believed false
Believed or known true
– Monotonocity
CSC411 Artificial Intelligence 5
Is-consistent-with Operator M
When reason, make sure the premises are
consistent
Format: M p – p is consistent with KB
Consider
– X good_student(X)  M study_hard(X) 
graduates(X)
– For all X who is a good student, if the fact that
X studies hard is consistent with KB, then X
will graduate
– Not necessary to prove that X study hard.
How to decide p is consistent with KB
– Negation as failure
– Heuristic-based and limited search
CSC411 Artificial Intelligence 6
Default Logic
Introduce a new format of inference rules:
– A(Z)  :B(Z)  C(Z)
– If A(Z) is provable, and it is consistent with
what we know to assume B(Z), then conclude
C(Z)
Compare with is-consistent-with operator
– Similar
– Difference is the reasoning method
In default logic, new rules are used to infer sets of
plausible extensions
– Example:
X good_student(X)  :study_hard(X) 
graduates(X)
Y party(Y)  :not(study_hard(Y)) 
not(graduates(X))
CSC411 Artificial Intelligence 7
Stanford Certainty Factor Algebra
Measure of confidence or believe
Summation may not be 1
Simple case:
– Confidence for: MB(H|E)
– Confidence against: MD(H|E)
– Properties:
1>MB(H|E)>0 while MD(H|E)=0, or
1>MD(H|E)>0 while MB(H|E)=0
– Put together
CF(H|E) = MB(H|E) – MD(H|E)
1 > CF(H|E) > -1
CSC411 Artificial Intelligence 8
CF Combination
Premises combination
– CF( P and Q) = min(CF(P), CF(Q))
– CF( P or Q) = max(CF(P), CF(Q))
Rule CF: each rule has a confidence measure
CF propagation
– Rule R: P  Q with CF=CF(R)
– CF(Q) = CF(P)CF(R)
Rule combination
– Rules R1: P1  Q: CF1(Q) = CF(P1)xCF(R1)
– R2: P2  Q: CF2(Q) = CF(P2)xCF(R2)
– CF(Q) =
CF1+CF2 – (CF1xCF2) if both positive
CF1+CF2 + (CF1xCF2) if both negative
(CF1+CF2)/(1-min(|CF1|,|CF2|)) otherwise
CSC411 Artificial Intelligence 9
Fuzzy Sets
Classic sets
– Completeness: x in either A or ¬A
– Exclusive: can not be in both A and ¬A
Fuzzy sets
– Violate the two assumptions
– Possibility theory -- measure of confidence or
believe
– Probability theory – randomness
– Process imprecision
– Introduce membership function
– Believe xA in some degree between 0 and 1,
inclusive
CSC411 Artificial Intelligence 10
The fuzzy set representation for “small integers.”
CSC411 Artificial Intelligence 11
A fuzzy set representation for the sets short, medium,
and tall males.
CSC411 Artificial Intelligence 12
Fuzzy Set Operations
Fuzzy set operations are defined as the
operations of membership functions
Complement: ¬A = C
– mC = 1 – mA
Union: A  B =C
– mC = max(mA, mB)
Intersection: A  B = C
– mC = min(mA, mB)
Difference: A – B = C
– mC = max(0, mA-mB)
CSC411 Artificial Intelligence 13
Fuzzy Inference Rules
Rule format and computation
– If x is A and y is B then z is C
mC(z) = min(mA(x), mB(y))
– If x is A or y is B then z is C
mC(z) = max(mA(x), mB(y))
– If x is not A then z is C
mC(z) = 1 – mA(x)
CSC411 Artificial Intelligence 14
The inverted pendulum and the angle θ and dθ/dt
input values.
CSC411 Artificial Intelligence 15
The fuzzy regions for the input values θ (a) and dθ/dt (b).
N – Negative, Z – Zero, P – Positive
CSC411 Artificial Intelligence 16
The fuzzy regions of the output value u, indicating the
movement of the pendulum base: Negative Big,
Negative, Zero, Positive, Positive Big.
CSC411 Artificial Intelligence 17
The fuzzificzation of the input measures
X1 = 1: mZ(X1) = mP(X1) = 0.5, mN(X1) = 0
X2 = -4: mZ(X2) = 0.2, mN(X2) = 0.8 , mP(X2) = 0
CSC411 Artificial Intelligence 18
The Fuzzy Associative
Matrix (FAM) for the
pendulum problem. The
input values are on the
left and top.
Fuzzy Rules:
CSC411 Artificial Intelligence 19
The fuzzy consequents (a) and their union (b). The
centroid of the union (-2) is the crisp output.
CSC411 Artificial Intelligence 20
Dempster-Shafer Theory
Probability theory limitation
– Assign a single number to measure any situation, no matter how
it is complex
– Cannot deal with missing evidence, heuristics, and limited
knowledge
Dempster-Shafer theory
– Extend probability theory
– Consider a set of propositions as a whole
– Assign a set of propositions an interval [believe, plausibility] to
constraint the degree of belief for each individual propositions in
the set
– The belief measure bel is in [0,1]
0 – no support evidence for a set of propositions
1 – full support evidence for a set of propositions
– The plausibility of p,
pl(p) = 1 – bel(not(p))
Reflect how evidence of not(p) relates to the possibility for belief in p
Bel(not(p))=1: full support for not(p), no possibility for p
Bel(not(p))=0: no support for not(p), full possibility for p
Range is also in [0,1]
CSC411 Artificial Intelligence 21
Properties of Dempster-Shafer
Initially, no support evidence for either
competing hypotheses, say h1 and h2
– Dempster-Shafer: [bel, pl] = [0, 1]
– Probability theory: p(h1)=p(h2)=0.5
Dempster-Shafer belief functions satisfy
weaker axioms than probability function
Two fundamental ideas:
– Obtaining belief degrees for one question from
subjective probabilities for related questions
– Using Dempster rule to combine these belief
degrees when they are based on independent
evidence
CSC411 Artificial Intelligence 22
An Example
Two persons M and B with reliabilities detect a computer and
claim the result independently. How you believe their claims?
Question (Q): detection claim
Related question (RQ): detectors’ reliability
Dempster-Shafer approach
– Obtain belief degrees for Q from subjective (prior) probabilities for RQ
for each person
– Combine belief degrees from two persons
Person M:
– reliability 0.9, unreliability 0.1
– Claim h1
– Belief degree of h1 is bel(h1)=0.9
– Belief degree of not(h1) is bel(not(h1))=0.0, different from probability
theory, since no evidence supporting not(h1)
– pl(h1) = 1 – bel(not(h1)) = 1-0 =1
– Thus belief measure for M claim h1 is [0.9, 1]
Person B:
– Reliability 0.8, unreliability 0.2
– Claim h2
– bel(h2) =0.8, bel(not(h2))=0, pl(h2)=1-bel(not(h2))=1-0
– Belief measure for B claim h2 is [0.8,1]
CSC411 Artificial Intelligence 23
Combining Belief Measure
Set of propositions: M claim h1 and B claim h2
– Case 1: h1 = h2
Reliability M and B: 09x0.8=0.72
Unreliability M and B: 0.1x0.2=0.02
The probability that at least one of two is reliable: 1-0.02=0.98
Belief measure for h1=h2 is [0.98,1]
– Case 2: h1 = not(h2)
Cannot be both correct and reliable
At least one is unreliable
– Reliable M and unreliable B: 0.9x(1-0.8)=0.18
– Reliable B and unreliable M: 0.8x(1-0.1)=0.08
– Unreliable M and B: (1-0.9)x(1-0.8)=0.02
– At least one is unreliable: 0.18+0.08+0.02=0.28
Given at least one is unreliable, posterior probabilities
– Reliable M and unreliable B: 0.18/0.28=0.643
– Reliable B and unreliable M: 0.08/0.28=0.286
Belief measure for h1
– Bel(h1)=0.643, bel(not(h1))=bel(h2)=0.286
– Pl(h1)=1-bel(not(h1))=1-0.286=0.714
– Belief measure: [0.643, 0.714]
Belief measure for h2
– Bel(h2)=0.286, bel(not(h2))=bel(h1)=0.683
– Pl(h2)=1-bel(not(h2))=1-0.683=0.317
– Belief measure: [0.286, 0.317]
CSC411 Artificial Intelligence 24
Dempster’s Rule
Assumption:
– probable questions are independent a priori
– As new evidence collected and conflicts, independency
may disappear
Two steps
1. Sort the uncertainties into a priori independent pieces of
evidence
2. Carry out Dempster rule
Consider the previous example
– After M and B claimed, a repair person is called to
check the computer, and both M and B witnessed this.
– Three independent items of evidence must be
combined
Not all evidence is directly supportive of
individual elements of a set of hypotheses, but
often supports different subsets of hypotheses,
in favor of some and against others
CSC411 Artificial Intelligence 25
General Dempster’s Rule
Q – an exhaustive set of mutually exclusive
hypotheses
Z – a subset of Q
M – probability density function to assign a
belief measure to Z
Mn(Z) – belief degree to Z, where n is the
number of sources of evidences
CSC411 Artificial Intelligence 26
Bayesian Belief Network
A computational model for reasoning to the best
explanation of a data set in the uncertainty
context
Motivation
– Reduce the number of parameters of the full Bayesian
model
– Show how the data can partition and focus reasoning
– Avoid use of a large joint probability table to compute
probabilities for all possible events combination
Assumption
– Events are either conditionally independent or their
correlations are so small that they can be ignored
Directed Graphical Model
– The events and (cause-effect) relationships form a
directed graph, where events are vertices and
relationships are links
CSC411 Artificial Intelligence 27
The Bayesian representation of the traffic problem with potential
explanations.
The joint probability distribution for the traffic and construction
variables
The Traffic Problem
Given bad traffic, what is the probability of road construction?
p(C|T)=p(C=t, T=t)/(p(C=t, T=t)+p(C=f, T=t))=.3/(.3+.1)=.75
CSC411 Artificial Intelligence 28
An Example
Traffic problem
– Events:
Road construction C
Accident A
Orange barrels B
Bad traffic T
Flashing lights L
– Joint probability
P(C,A,B,T,L)=p(C)*p(A|C)*p(B|C,A)*p(T|C,A,B)*p(L|C,A,B,T)
Number of parameters: 2^5=32
– Reduction
Assumption: Parameters are only dependent on parents
Calculation of joint probability
– P(C,A,B,T,L)=p(C)*p(A)*p(B|C)*p(T|C,A)*p(L|A)
– Number of parameters: 2+2+4+8+4=20
CSC411 Artificial Intelligence 29
BBN Definition
Links represent conditional probabilities for causal influence
These influences are directed: presence of some event
causes other events
These influences are not circular
Thus a BBN is a DAG: Directed Acyclic Graph
CSC411 Artificial Intelligence 30
Discrete Markov Process
Finite state machine
– A graphical representation
– State transition depends on input stream
– States and transitions reflect properties of a
formal language
Probabilistic finite state machine
– A finite state machine
– Transition function represented by a
probability distribution on the current state
Discrete Markov process (chain, machine)
– A specialization of probabilistic finite state
machine
– Ignores its input values
CSC411 Artificial Intelligence 31
A Markov state machine or Markov chain with four states, s1,
..., s4
At any time the system is in one of distinct states
The system undergoes state change or remain
Divide time into discrete intervals: t1, t2, …, tn
Change state according to the probability distribution of
each state
S(t) – the actual state at time t
p(S(t)) = p(S(t)|S(t-1), s(t-2), s(t-3), …)
First-order markov chain
– Only depends on the direct predecessor state
– P(S(t)) = p(S(t)|S(t-1))
CSC411 Artificial Intelligence 32
Observable Markov Model
Assume p(S(t)|S(t-1)) is time invariant, that is, transition
between specific states retains the same probabilistic
relationship
State transition probability aij between si and sj:
– aij=p(S(t)=si|S(t-1)=sj), 1<=i,j<=N
– If i=j, no transition (remain the same state)
– Properties: aij >=0, iaij=1
CSC411 Artificial Intelligence 33
S1 – sun
S2 – cloudy
S3 – fog
S4 – precipitation
Time intervals:
noon to noon
Question: suppose that
today is sunny, what is
the probability of the
next five days being
sunny, sunny, cloudy,
cloudy, precipitation?

More Related Content

Similar to Chapter 9

Statistical Machine________ Learning.ppt
Statistical Machine________ Learning.pptStatistical Machine________ Learning.ppt
Statistical Machine________ Learning.ppt
SandeepGupta229023
 
Into to prob_prog_hari (2)
Into to prob_prog_hari (2)Into to prob_prog_hari (2)
Into to prob_prog_hari (2)
Hariharan Chandrasekaran
 
. An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic .... An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic ...
butest
 
Logics of Context and Modal Type Theories
Logics of Context and Modal Type TheoriesLogics of Context and Modal Type Theories
Logics of Context and Modal Type Theories
Valeria de Paiva
 
Low-rank matrix approximations in Python by Christian Thurau PyData 2014
Low-rank matrix approximations in Python by Christian Thurau PyData 2014Low-rank matrix approximations in Python by Christian Thurau PyData 2014
Low-rank matrix approximations in Python by Christian Thurau PyData 2014
PyData
 
Resources
ResourcesResources
Resources
butest
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
kkkc
 
Machine Learning Algorithms Review(Part 2)
Machine Learning Algorithms Review(Part 2)Machine Learning Algorithms Review(Part 2)
Machine Learning Algorithms Review(Part 2)
Zihui Li
 
Class9_PCA_final.ppt
Class9_PCA_final.pptClass9_PCA_final.ppt
Class9_PCA_final.ppt
MaTruongThanh002937
 
2.7 other classifiers
2.7 other classifiers2.7 other classifiers
2.7 other classifiers
Krish_ver2
 
advanced_statistics.pdf
advanced_statistics.pdfadvanced_statistics.pdf
advanced_statistics.pdf
GerryMakilan2
 
Chapter4
Chapter4Chapter4
Chapter4
Vu Vo
 
CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)
Mostafa G. M. Mostafa
 
Representing uncertainty in expert systems
Representing uncertainty in expert systemsRepresenting uncertainty in expert systems
Representing uncertainty in expert systems
bhupendra kumar
 
Unit-2.ppt
Unit-2.pptUnit-2.ppt
Unit-2.ppt
AshwaniShukla47
 
Ppt on artifishail intelligence
Ppt on artifishail intelligencePpt on artifishail intelligence
Ppt on artifishail intelligence
snehal_gongle
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)
Mostafa G. M. Mostafa
 
Lecture 2 Basic Concepts in Machine Learning for Language Technology
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyLecture 2 Basic Concepts in Machine Learning for Language Technology
Lecture 2 Basic Concepts in Machine Learning for Language Technology
Marina Santini
 
Bayesian Neural Networks
Bayesian Neural NetworksBayesian Neural Networks
Bayesian Neural Networks
Natan Katz
 
Intro to Model Selection
Intro to Model SelectionIntro to Model Selection
Intro to Model Selection
chenhm
 

Similar to Chapter 9 (20)

Statistical Machine________ Learning.ppt
Statistical Machine________ Learning.pptStatistical Machine________ Learning.ppt
Statistical Machine________ Learning.ppt
 
Into to prob_prog_hari (2)
Into to prob_prog_hari (2)Into to prob_prog_hari (2)
Into to prob_prog_hari (2)
 
. An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic .... An introduction to machine learning and probabilistic ...
. An introduction to machine learning and probabilistic ...
 
Logics of Context and Modal Type Theories
Logics of Context and Modal Type TheoriesLogics of Context and Modal Type Theories
Logics of Context and Modal Type Theories
 
Low-rank matrix approximations in Python by Christian Thurau PyData 2014
Low-rank matrix approximations in Python by Christian Thurau PyData 2014Low-rank matrix approximations in Python by Christian Thurau PyData 2014
Low-rank matrix approximations in Python by Christian Thurau PyData 2014
 
Resources
ResourcesResources
Resources
 
Introduction to Machine Learning
Introduction to Machine LearningIntroduction to Machine Learning
Introduction to Machine Learning
 
Machine Learning Algorithms Review(Part 2)
Machine Learning Algorithms Review(Part 2)Machine Learning Algorithms Review(Part 2)
Machine Learning Algorithms Review(Part 2)
 
Class9_PCA_final.ppt
Class9_PCA_final.pptClass9_PCA_final.ppt
Class9_PCA_final.ppt
 
2.7 other classifiers
2.7 other classifiers2.7 other classifiers
2.7 other classifiers
 
advanced_statistics.pdf
advanced_statistics.pdfadvanced_statistics.pdf
advanced_statistics.pdf
 
Chapter4
Chapter4Chapter4
Chapter4
 
CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)
 
Representing uncertainty in expert systems
Representing uncertainty in expert systemsRepresenting uncertainty in expert systems
Representing uncertainty in expert systems
 
Unit-2.ppt
Unit-2.pptUnit-2.ppt
Unit-2.ppt
 
Ppt on artifishail intelligence
Ppt on artifishail intelligencePpt on artifishail intelligence
Ppt on artifishail intelligence
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)
 
Lecture 2 Basic Concepts in Machine Learning for Language Technology
Lecture 2 Basic Concepts in Machine Learning for Language TechnologyLecture 2 Basic Concepts in Machine Learning for Language Technology
Lecture 2 Basic Concepts in Machine Learning for Language Technology
 
Bayesian Neural Networks
Bayesian Neural NetworksBayesian Neural Networks
Bayesian Neural Networks
 
Intro to Model Selection
Intro to Model SelectionIntro to Model Selection
Intro to Model Selection
 

Recently uploaded

Generative AI Use cases applications solutions and implementation.pdf
Generative AI Use cases applications solutions and implementation.pdfGenerative AI Use cases applications solutions and implementation.pdf
Generative AI Use cases applications solutions and implementation.pdf
mahaffeycheryld
 
Gas agency management system project report.pdf
Gas agency management system project report.pdfGas agency management system project report.pdf
Gas agency management system project report.pdf
Kamal Acharya
 
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student MemberIEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
VICTOR MAESTRE RAMIREZ
 
Null Bangalore | Pentesters Approach to AWS IAM
Null Bangalore | Pentesters Approach to AWS IAMNull Bangalore | Pentesters Approach to AWS IAM
Null Bangalore | Pentesters Approach to AWS IAM
Divyanshu
 
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf
Yasser Mahgoub
 
ITSM Integration with MuleSoft.pptx
ITSM  Integration with MuleSoft.pptxITSM  Integration with MuleSoft.pptx
ITSM Integration with MuleSoft.pptx
VANDANAMOHANGOUDA
 
Embedded machine learning-based road conditions and driving behavior monitoring
Embedded machine learning-based road conditions and driving behavior monitoringEmbedded machine learning-based road conditions and driving behavior monitoring
Embedded machine learning-based road conditions and driving behavior monitoring
IJECEIAES
 
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024
Sinan KOZAK
 
Properties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptxProperties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptx
MDSABBIROJJAMANPAYEL
 
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...
PriyankaKilaniya
 
Welding Metallurgy Ferrous Materials.pdf
Welding Metallurgy Ferrous Materials.pdfWelding Metallurgy Ferrous Materials.pdf
Welding Metallurgy Ferrous Materials.pdf
AjmalKhan50578
 
Applications of artificial Intelligence in Mechanical Engineering.pdf
Applications of artificial Intelligence in Mechanical Engineering.pdfApplications of artificial Intelligence in Mechanical Engineering.pdf
Applications of artificial Intelligence in Mechanical Engineering.pdf
Atif Razi
 
一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理
一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理
一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理
nedcocy
 
4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf
4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf
4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf
Gino153088
 
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...
shadow0702a
 
Design and optimization of ion propulsion drone
Design and optimization of ion propulsion droneDesign and optimization of ion propulsion drone
Design and optimization of ion propulsion drone
bjmsejournal
 
Comparative analysis between traditional aquaponics and reconstructed aquapon...
Comparative analysis between traditional aquaponics and reconstructed aquapon...Comparative analysis between traditional aquaponics and reconstructed aquapon...
Comparative analysis between traditional aquaponics and reconstructed aquapon...
bijceesjournal
 
22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt
KrishnaveniKrishnara1
 
1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf
1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf
1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf
MadhavJungKarki
 
Data Driven Maintenance | UReason Webinar
Data Driven Maintenance | UReason WebinarData Driven Maintenance | UReason Webinar
Data Driven Maintenance | UReason Webinar
UReason
 

Recently uploaded (20)

Generative AI Use cases applications solutions and implementation.pdf
Generative AI Use cases applications solutions and implementation.pdfGenerative AI Use cases applications solutions and implementation.pdf
Generative AI Use cases applications solutions and implementation.pdf
 
Gas agency management system project report.pdf
Gas agency management system project report.pdfGas agency management system project report.pdf
Gas agency management system project report.pdf
 
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student MemberIEEE Aerospace and Electronic Systems Society as a Graduate Student Member
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
 
Null Bangalore | Pentesters Approach to AWS IAM
Null Bangalore | Pentesters Approach to AWS IAMNull Bangalore | Pentesters Approach to AWS IAM
Null Bangalore | Pentesters Approach to AWS IAM
 
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf
2008 BUILDING CONSTRUCTION Illustrated - Ching Chapter 08 Doors and Windows.pdf
 
ITSM Integration with MuleSoft.pptx
ITSM  Integration with MuleSoft.pptxITSM  Integration with MuleSoft.pptx
ITSM Integration with MuleSoft.pptx
 
Embedded machine learning-based road conditions and driving behavior monitoring
Embedded machine learning-based road conditions and driving behavior monitoringEmbedded machine learning-based road conditions and driving behavior monitoring
Embedded machine learning-based road conditions and driving behavior monitoring
 
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024
 
Properties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptxProperties Railway Sleepers and Test.pptx
Properties Railway Sleepers and Test.pptx
 
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...
Prediction of Electrical Energy Efficiency Using Information on Consumer's Ac...
 
Welding Metallurgy Ferrous Materials.pdf
Welding Metallurgy Ferrous Materials.pdfWelding Metallurgy Ferrous Materials.pdf
Welding Metallurgy Ferrous Materials.pdf
 
Applications of artificial Intelligence in Mechanical Engineering.pdf
Applications of artificial Intelligence in Mechanical Engineering.pdfApplications of artificial Intelligence in Mechanical Engineering.pdf
Applications of artificial Intelligence in Mechanical Engineering.pdf
 
一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理
一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理
一比一原版(爱大毕业证书)爱荷华大学毕业证如何办理
 
4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf
4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf
4. Mosca vol I -Fisica-Tipler-5ta-Edicion-Vol-1.pdf
 
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...
Use PyCharm for remote debugging of WSL on a Windo cf5c162d672e4e58b4dde5d797...
 
Design and optimization of ion propulsion drone
Design and optimization of ion propulsion droneDesign and optimization of ion propulsion drone
Design and optimization of ion propulsion drone
 
Comparative analysis between traditional aquaponics and reconstructed aquapon...
Comparative analysis between traditional aquaponics and reconstructed aquapon...Comparative analysis between traditional aquaponics and reconstructed aquapon...
Comparative analysis between traditional aquaponics and reconstructed aquapon...
 
22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt22CYT12-Unit-V-E Waste and its Management.ppt
22CYT12-Unit-V-E Waste and its Management.ppt
 
1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf
1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf
1FIDIC-CONSTRUCTION-CONTRACT-2ND-ED-2017-RED-BOOK.pdf
 
Data Driven Maintenance | UReason Webinar
Data Driven Maintenance | UReason WebinarData Driven Maintenance | UReason Webinar
Data Driven Maintenance | UReason Webinar
 

Chapter 9

  • 1. CSC411 Artificial Intelligence 1 Chapter 9 Reasoning in Uncertain Situations Contents Uncertain situations Non-monotonic logic and reasoning Certainty Factor algebra Fuzzy logic and reasoning Dempster-Shafer theory of evidence Bayesian belief network Markov models
  • 2. CSC411 Artificial Intelligence 2 Traditional Logic Based on predicate logic Three important assumptions: – Predicate descriptions are sufficient w.r.t. to the domain – Information is consistent – Knowledge base grows monotonically
  • 3. CSC411 Artificial Intelligence 3 Non-monotonic Logic Addresses the three assumptions of traditional logic – Knowledge is incomplete No knowledge about p: true or false? Prolog – closed world assumption – Knowledge is inconsistent Based on how the world usually works Most birds fly, but Ostrich doesn’t – Knowledge base grows non-monotonically New observation may contradict the existing knowledge, thus the existing knowledge may need removal. Inference based on assumptions, how come if the assumptions are later shown to be incorrect Three modal operators are introduced
  • 4. CSC411 Artificial Intelligence 4 Unless Operator New information may invalidate previous results Implemented in TMS – Truth Maintenance Systems to keep track of the reasoning steps and preserve the KB consistency Introduce Unless operator – Support inferences based on the belief that its argument is not true – Consider p(X) unless q(X)  r(X) If p(X) is true and not believe q(X) true then r(X) p(Z) r(W)  s(W) From above, conclude s(X). Later, change believe or find q(X) true, what happens? Retract r(X) and s(X) – Unless deals with believe, not truth Either unknown or believed false Believed or known true – Monotonocity
  • 5. CSC411 Artificial Intelligence 5 Is-consistent-with Operator M When reason, make sure the premises are consistent Format: M p – p is consistent with KB Consider – X good_student(X)  M study_hard(X)  graduates(X) – For all X who is a good student, if the fact that X studies hard is consistent with KB, then X will graduate – Not necessary to prove that X study hard. How to decide p is consistent with KB – Negation as failure – Heuristic-based and limited search
  • 6. CSC411 Artificial Intelligence 6 Default Logic Introduce a new format of inference rules: – A(Z)  :B(Z)  C(Z) – If A(Z) is provable, and it is consistent with what we know to assume B(Z), then conclude C(Z) Compare with is-consistent-with operator – Similar – Difference is the reasoning method In default logic, new rules are used to infer sets of plausible extensions – Example: X good_student(X)  :study_hard(X)  graduates(X) Y party(Y)  :not(study_hard(Y))  not(graduates(X))
  • 7. CSC411 Artificial Intelligence 7 Stanford Certainty Factor Algebra Measure of confidence or believe Summation may not be 1 Simple case: – Confidence for: MB(H|E) – Confidence against: MD(H|E) – Properties: 1>MB(H|E)>0 while MD(H|E)=0, or 1>MD(H|E)>0 while MB(H|E)=0 – Put together CF(H|E) = MB(H|E) – MD(H|E) 1 > CF(H|E) > -1
  • 8. CSC411 Artificial Intelligence 8 CF Combination Premises combination – CF( P and Q) = min(CF(P), CF(Q)) – CF( P or Q) = max(CF(P), CF(Q)) Rule CF: each rule has a confidence measure CF propagation – Rule R: P  Q with CF=CF(R) – CF(Q) = CF(P)CF(R) Rule combination – Rules R1: P1  Q: CF1(Q) = CF(P1)xCF(R1) – R2: P2  Q: CF2(Q) = CF(P2)xCF(R2) – CF(Q) = CF1+CF2 – (CF1xCF2) if both positive CF1+CF2 + (CF1xCF2) if both negative (CF1+CF2)/(1-min(|CF1|,|CF2|)) otherwise
  • 9. CSC411 Artificial Intelligence 9 Fuzzy Sets Classic sets – Completeness: x in either A or ¬A – Exclusive: can not be in both A and ¬A Fuzzy sets – Violate the two assumptions – Possibility theory -- measure of confidence or believe – Probability theory – randomness – Process imprecision – Introduce membership function – Believe xA in some degree between 0 and 1, inclusive
  • 10. CSC411 Artificial Intelligence 10 The fuzzy set representation for “small integers.”
  • 11. CSC411 Artificial Intelligence 11 A fuzzy set representation for the sets short, medium, and tall males.
  • 12. CSC411 Artificial Intelligence 12 Fuzzy Set Operations Fuzzy set operations are defined as the operations of membership functions Complement: ¬A = C – mC = 1 – mA Union: A  B =C – mC = max(mA, mB) Intersection: A  B = C – mC = min(mA, mB) Difference: A – B = C – mC = max(0, mA-mB)
  • 13. CSC411 Artificial Intelligence 13 Fuzzy Inference Rules Rule format and computation – If x is A and y is B then z is C mC(z) = min(mA(x), mB(y)) – If x is A or y is B then z is C mC(z) = max(mA(x), mB(y)) – If x is not A then z is C mC(z) = 1 – mA(x)
  • 14. CSC411 Artificial Intelligence 14 The inverted pendulum and the angle θ and dθ/dt input values.
  • 15. CSC411 Artificial Intelligence 15 The fuzzy regions for the input values θ (a) and dθ/dt (b). N – Negative, Z – Zero, P – Positive
  • 16. CSC411 Artificial Intelligence 16 The fuzzy regions of the output value u, indicating the movement of the pendulum base: Negative Big, Negative, Zero, Positive, Positive Big.
  • 17. CSC411 Artificial Intelligence 17 The fuzzificzation of the input measures X1 = 1: mZ(X1) = mP(X1) = 0.5, mN(X1) = 0 X2 = -4: mZ(X2) = 0.2, mN(X2) = 0.8 , mP(X2) = 0
  • 18. CSC411 Artificial Intelligence 18 The Fuzzy Associative Matrix (FAM) for the pendulum problem. The input values are on the left and top. Fuzzy Rules:
  • 19. CSC411 Artificial Intelligence 19 The fuzzy consequents (a) and their union (b). The centroid of the union (-2) is the crisp output.
  • 20. CSC411 Artificial Intelligence 20 Dempster-Shafer Theory Probability theory limitation – Assign a single number to measure any situation, no matter how it is complex – Cannot deal with missing evidence, heuristics, and limited knowledge Dempster-Shafer theory – Extend probability theory – Consider a set of propositions as a whole – Assign a set of propositions an interval [believe, plausibility] to constraint the degree of belief for each individual propositions in the set – The belief measure bel is in [0,1] 0 – no support evidence for a set of propositions 1 – full support evidence for a set of propositions – The plausibility of p, pl(p) = 1 – bel(not(p)) Reflect how evidence of not(p) relates to the possibility for belief in p Bel(not(p))=1: full support for not(p), no possibility for p Bel(not(p))=0: no support for not(p), full possibility for p Range is also in [0,1]
  • 21. CSC411 Artificial Intelligence 21 Properties of Dempster-Shafer Initially, no support evidence for either competing hypotheses, say h1 and h2 – Dempster-Shafer: [bel, pl] = [0, 1] – Probability theory: p(h1)=p(h2)=0.5 Dempster-Shafer belief functions satisfy weaker axioms than probability function Two fundamental ideas: – Obtaining belief degrees for one question from subjective probabilities for related questions – Using Dempster rule to combine these belief degrees when they are based on independent evidence
  • 22. CSC411 Artificial Intelligence 22 An Example Two persons M and B with reliabilities detect a computer and claim the result independently. How you believe their claims? Question (Q): detection claim Related question (RQ): detectors’ reliability Dempster-Shafer approach – Obtain belief degrees for Q from subjective (prior) probabilities for RQ for each person – Combine belief degrees from two persons Person M: – reliability 0.9, unreliability 0.1 – Claim h1 – Belief degree of h1 is bel(h1)=0.9 – Belief degree of not(h1) is bel(not(h1))=0.0, different from probability theory, since no evidence supporting not(h1) – pl(h1) = 1 – bel(not(h1)) = 1-0 =1 – Thus belief measure for M claim h1 is [0.9, 1] Person B: – Reliability 0.8, unreliability 0.2 – Claim h2 – bel(h2) =0.8, bel(not(h2))=0, pl(h2)=1-bel(not(h2))=1-0 – Belief measure for B claim h2 is [0.8,1]
  • 23. CSC411 Artificial Intelligence 23 Combining Belief Measure Set of propositions: M claim h1 and B claim h2 – Case 1: h1 = h2 Reliability M and B: 09x0.8=0.72 Unreliability M and B: 0.1x0.2=0.02 The probability that at least one of two is reliable: 1-0.02=0.98 Belief measure for h1=h2 is [0.98,1] – Case 2: h1 = not(h2) Cannot be both correct and reliable At least one is unreliable – Reliable M and unreliable B: 0.9x(1-0.8)=0.18 – Reliable B and unreliable M: 0.8x(1-0.1)=0.08 – Unreliable M and B: (1-0.9)x(1-0.8)=0.02 – At least one is unreliable: 0.18+0.08+0.02=0.28 Given at least one is unreliable, posterior probabilities – Reliable M and unreliable B: 0.18/0.28=0.643 – Reliable B and unreliable M: 0.08/0.28=0.286 Belief measure for h1 – Bel(h1)=0.643, bel(not(h1))=bel(h2)=0.286 – Pl(h1)=1-bel(not(h1))=1-0.286=0.714 – Belief measure: [0.643, 0.714] Belief measure for h2 – Bel(h2)=0.286, bel(not(h2))=bel(h1)=0.683 – Pl(h2)=1-bel(not(h2))=1-0.683=0.317 – Belief measure: [0.286, 0.317]
  • 24. CSC411 Artificial Intelligence 24 Dempster’s Rule Assumption: – probable questions are independent a priori – As new evidence collected and conflicts, independency may disappear Two steps 1. Sort the uncertainties into a priori independent pieces of evidence 2. Carry out Dempster rule Consider the previous example – After M and B claimed, a repair person is called to check the computer, and both M and B witnessed this. – Three independent items of evidence must be combined Not all evidence is directly supportive of individual elements of a set of hypotheses, but often supports different subsets of hypotheses, in favor of some and against others
  • 25. CSC411 Artificial Intelligence 25 General Dempster’s Rule Q – an exhaustive set of mutually exclusive hypotheses Z – a subset of Q M – probability density function to assign a belief measure to Z Mn(Z) – belief degree to Z, where n is the number of sources of evidences
  • 26. CSC411 Artificial Intelligence 26 Bayesian Belief Network A computational model for reasoning to the best explanation of a data set in the uncertainty context Motivation – Reduce the number of parameters of the full Bayesian model – Show how the data can partition and focus reasoning – Avoid use of a large joint probability table to compute probabilities for all possible events combination Assumption – Events are either conditionally independent or their correlations are so small that they can be ignored Directed Graphical Model – The events and (cause-effect) relationships form a directed graph, where events are vertices and relationships are links
  • 27. CSC411 Artificial Intelligence 27 The Bayesian representation of the traffic problem with potential explanations. The joint probability distribution for the traffic and construction variables The Traffic Problem Given bad traffic, what is the probability of road construction? p(C|T)=p(C=t, T=t)/(p(C=t, T=t)+p(C=f, T=t))=.3/(.3+.1)=.75
  • 28. CSC411 Artificial Intelligence 28 An Example Traffic problem – Events: Road construction C Accident A Orange barrels B Bad traffic T Flashing lights L – Joint probability P(C,A,B,T,L)=p(C)*p(A|C)*p(B|C,A)*p(T|C,A,B)*p(L|C,A,B,T) Number of parameters: 2^5=32 – Reduction Assumption: Parameters are only dependent on parents Calculation of joint probability – P(C,A,B,T,L)=p(C)*p(A)*p(B|C)*p(T|C,A)*p(L|A) – Number of parameters: 2+2+4+8+4=20
  • 29. CSC411 Artificial Intelligence 29 BBN Definition Links represent conditional probabilities for causal influence These influences are directed: presence of some event causes other events These influences are not circular Thus a BBN is a DAG: Directed Acyclic Graph
  • 30. CSC411 Artificial Intelligence 30 Discrete Markov Process Finite state machine – A graphical representation – State transition depends on input stream – States and transitions reflect properties of a formal language Probabilistic finite state machine – A finite state machine – Transition function represented by a probability distribution on the current state Discrete Markov process (chain, machine) – A specialization of probabilistic finite state machine – Ignores its input values
  • 31. CSC411 Artificial Intelligence 31 A Markov state machine or Markov chain with four states, s1, ..., s4 At any time the system is in one of distinct states The system undergoes state change or remain Divide time into discrete intervals: t1, t2, …, tn Change state according to the probability distribution of each state S(t) – the actual state at time t p(S(t)) = p(S(t)|S(t-1), s(t-2), s(t-3), …) First-order markov chain – Only depends on the direct predecessor state – P(S(t)) = p(S(t)|S(t-1))
  • 32. CSC411 Artificial Intelligence 32 Observable Markov Model Assume p(S(t)|S(t-1)) is time invariant, that is, transition between specific states retains the same probabilistic relationship State transition probability aij between si and sj: – aij=p(S(t)=si|S(t-1)=sj), 1<=i,j<=N – If i=j, no transition (remain the same state) – Properties: aij >=0, iaij=1
  • 33. CSC411 Artificial Intelligence 33 S1 – sun S2 – cloudy S3 – fog S4 – precipitation Time intervals: noon to noon Question: suppose that today is sunny, what is the probability of the next five days being sunny, sunny, cloudy, cloudy, precipitation?