SlideShare a Scribd company logo
1 of 20
USING PRIOR KNOWLEDGE TO
INITIALIZE THE HYPOTHESIS
Swapna,.C
3 USING PRIOR KNOWLEDGE TO
INITIALIZE THE HYPOTHESIS
• One approach to using prior knowledge is to initialize
the hypothesis to perfectly fit the domain theory, then
inductively refine this initial hypothesis as needed to fit
the training data. This approach is used by the KBANN
(Knowledge-Based Artificial Neural Network)
algorithm to learn artificial neural networks.
• In KBANN an initial network is first constructed so
that for every possible instance, the classification
assigned by the network is identical to that assigned by
the domain theory. The BACKPROPAGATION Algorithm
is then employed to adjust the weights of this initial
network as needed to fit the training examples.
• In KBANN , if the domain theory is only approximately
correct, initializing the network to fit this domain
theory will give a better starting approximation to the
target function than initializing the network to random
initial weights. This should lead, in turn, to better
generalization accuracy for the final hypothesis.
• Initialize-the-hypothesis approach to using the
domain theory has been explored by several
researchers, including Shavlik and Towel1 (1989),
Towel1 and Shavlik (1994), Fu (1989, 1993), and Pratt
(1993a, 1993b). We will use the KBANN algorithm
described in Shavlik and Towel1 (1989) to illustrate this
approach.
3.1 The KBANN Algorithm
• The KBANN algorithm exemplifies the
initialize-the-hypothesis approach to using
domain theories. It assumes a domain theory
represented by a set of propositional,
nonrecursive Horn clauses. A Horn clause is
propositional if it contains no variables.
• Given:
• A set of training examples
• A domain theory consisting of nonrecursive,
propositional Horn clauses
• Determine:
• An artificial neural network that fits the training
examples, biased by the domain theory.
• The two stages of the KBANN algorithm are first
to create an artificial neural network that
perfectly fits the domain theory and second to
use the BACKPROPACATION algorithm to refine
this initial network to fit the training examples.
3.2 An Illustrative Example
• Each instance describes a physical object in terms of
the material from which it is made, whether it is light,
etc. The task is to learn the target concept Cup defined
over such physical objects.
The domain theory defines a Cup as an object that is
Stable, Liftable, and an OpenVessel.
• The domain theory also defines each of these three
attributes in terms of more primitive attributes,
terminating in the primitive, operational attributes that
describe the instances. The domain theory is not
perfectly consistent with the training examples.
• For example, the domain theory fails to classify
the second and third training examples as
positive examples. Nevertheless, the domain
theory forms a useful approximation to the target
concept.
• KBANN uses the domain theory and training
examples together to learn the target concept
more accurately than it could from either alone.
• In the first stage of the KBANN algorithm (steps
1-3 in the algorithm), an initial network is
constructed that is consistent with the domain
theory.
•For example, the network constructed from the Cup domain theory is
shown in Figure. In general the network is constructed by creating a
sigmoid threshold unit for each Horn clause in the domain theory.
•KBANN follows the convention that a sigmoid output value greater than
0.5 is interpreted as True and a value below 0.5 as False.
Each unit is therefore constructed so that its output will be greater than
0.5 just in those cases where the corresponding Horn clause applies. For
each antecedent to the Horn clause, an input is created to the
corresponding sigmoid unit. The weights of the sigmoid unit are then set
so that it computes the logical AND of its inputs.
In particular, for each input corresponding to a non-negated antecedent,
the weight is set to some positive constant W. For each input
corresponding to a negated antecedent, the weight is set to - W.
The threshold weight of the unit, wo is then set to -(n- .5) W, where n is
the number of non-negated antecedents. When unit input values are
1 or 0, this assures that their weighted sum plus wo will be positive (and
the sigmoid output will therefore be greater than 0.5) if and only if all
clause antecedents are satisfied.
•KBANN algorithm can correctly encode the domain theory for
arbitrarily deep networks. Towell and Shavlik (1994) report using
W = 4.0 in many of their experiments.
•Each sigmoid unit input is connected to the appropriate network input
or to the output of another sigmoid unit, to mirror the graph of
dependencies among the corresponding attributes in the domain
theory. As a final step many additional inputs are added to each
threshold unit, with their weights set approximately to zero. The role of
these additional connections is to enable the network to inductively
learn additional dependencies beyond those suggested by the given
domain theory.
The solid lines in the network of Figure indicate unit inputs with
weights of W, whereas the lightly shaded lines indicate connections
with initial weights near zero. It is easy to verify that for sufficiently
large values of W this network will output values identical to the
predictions of the domain theory.
In the Cup example, however, the domain theory and training data are
inconsistent, and this step therefore alters the initial network weights. The
resulting trained network is summarized in Figure with dark solid lines
indicating the largest positive weights, dashed lines indicating the
largest negative weights, and light lines indicating negligible weights.
Although the initial network misclassifies several training examples
from Table , the refined network of Figure perfectly classifies all of
these training examples.
In Figure significant new dependencies were discovered during the
inductive step, including the dependency of the Liftable unit on the
feature MadeOfStyrofoam. It is important to keep in mind that while
the unit labeled Liftable was initially defined by the given Horn clause
for Liftable, the subsequent weight changes performed by
BACKPROPAGATION have dramatically changed, the meaning of this
hidden unit. After training of the network, this unit may take on a very
different meaning unrelated to the initial notion of Liftable.
Remarks:
•KBANN analytically creates a network equivalent to the given
domain theory, then inductively refines this initial hypothesis to
better fit the training data. In doing so, it modifies the network
weights as needed to overcome inconsistencies between the
domain theory and observed data.
•The chief benefit of KBANN over purely inductive
BACKPROPAGATION beginning with random initial weights) is that
it typically generalizes more accurately than BACKPROPAGATION
when given an approximately correct domain theory, especially
when training data is scarce..
For example, Towel1 et al. (1990) describe the application of KBANN to a
molecular genetics problem. Here the task was to learn to recognize DNA
segments called promoter regions, which influence gene activity.
In this experiment KBANN was given an initial domain theory obtained
from a molecular geneticist, and a set of 53 positive and 53 negative
training examples of promoter regions. Performance was evaluated using
a leave-one-out strategy in which the system was run 106 different times.
On each iteration KBANN was trained using 105 of the 106 examples and
tested on the remaining example.
The results of these 106 experiments were accumulated to provide an
estimate of the true error rate. KBANN obtained an error rate of 41106,
compared to an error rate of 81106 using standard BACKPROPAGATAION
variant of the KBANN approach was applied by Fu (1993), who reports an
error rate of 21106 on the same data. Thus, the impact of prior
knowledge in these experiments was to reduce significantly the error
rate. The training data for this experiment is available at World WideWeb
site http:llwww.ics.uci.edu/~mlearn/MLRepository.htm
To understand the significance of KBANN it is useful to consider how its
hypothesis search differs from that of the purely inductive BACKPROPAGATION
algorithm. The hypothesis space search conducted by both algorithms is
depicted schematically in Figure
The key difference is the initial hypothesis from which weight
tuning is performed. In the case that multiple hypotheses (weight
vectors) can be found that fit the data-a condition that will be
especially likely when training data is scarce-KBANN is likely to
converge to a hypothesis that generalizes beyond the data in a way
that is more similar to the domain theory predictions.
On the other hand, the particular hypothesis to which
BACKPROPAGATION converges will more likely be a hypothesis
with small weights, corresponding roughly to a generalization bias
of smoothly interpolating between training examples. In brief,
KBANN uses a domain-specific theory to bias generalization,
whereas BACKPROPAGATION uses a domain-independent syntactic
bias toward small weight values. Note in this summary we have
ignored the effect of local minima on the search.
Limitations of KBANN include the fact that it can accommodate only
propositional domain theories; that is, collections of variable-free Horn
clauses. It is also possible for KBANN to be misled when given highly
inaccurate domain theories, so that its generalization accuracy can
deteriorate below the level of BACKPROPAGATION.
KBANN illustrates the initialize-the-hypothesis approach to combining
analytical and inductive learning. These approaches vary in the exact
technique for constructing the initial network, the application of
BACKPROPAGATION to weight tuning, and in methods for extracting
symbolic descriptions from the refined network. Pratt (1993a, 1993b)
describes an initialize-the-hypothesis approach in which the prior
knowledge is provided by a previously learned neural network for a
related task, rather than a manually provided symbolic domain theory.
Here the prior knowledge corresponds to a set of conditional
independence assumptions that determine the graph structure of the
Bayes net, whose conditional probability tables are then induced from
the training data.

More Related Content

What's hot

Advanced topics in artificial neural networks
Advanced topics in artificial neural networksAdvanced topics in artificial neural networks
Advanced topics in artificial neural networksswapnac12
 
Evaluating hypothesis
Evaluating  hypothesisEvaluating  hypothesis
Evaluating hypothesisswapnac12
 
Target language in compiler design
Target language in compiler designTarget language in compiler design
Target language in compiler designMuhammad Haroon
 
Back propagation
Back propagationBack propagation
Back propagationNagarajan
 
MACHINE LEARNING-LEARNING RULE
MACHINE LEARNING-LEARNING RULEMACHINE LEARNING-LEARNING RULE
MACHINE LEARNING-LEARNING RULEDrBindhuM
 
Learning sets of rules, Sequential Learning Algorithm,FOIL
Learning sets of rules, Sequential Learning Algorithm,FOILLearning sets of rules, Sequential Learning Algorithm,FOIL
Learning sets of rules, Sequential Learning Algorithm,FOILPavithra Thippanaik
 
Issues in knowledge representation
Issues in knowledge representationIssues in knowledge representation
Issues in knowledge representationSravanthi Emani
 
Register allocation and assignment
Register allocation and assignmentRegister allocation and assignment
Register allocation and assignmentKarthi Keyan
 
Unification and Lifting
Unification and LiftingUnification and Lifting
Unification and LiftingMegha Sharma
 
Problem reduction AND OR GRAPH & AO* algorithm.ppt
Problem reduction AND OR GRAPH & AO* algorithm.pptProblem reduction AND OR GRAPH & AO* algorithm.ppt
Problem reduction AND OR GRAPH & AO* algorithm.pptarunsingh660
 
Knowledge representation and Predicate logic
Knowledge representation and Predicate logicKnowledge representation and Predicate logic
Knowledge representation and Predicate logicAmey Kerkar
 
Support Vector Machines ( SVM )
Support Vector Machines ( SVM ) Support Vector Machines ( SVM )
Support Vector Machines ( SVM ) Mohammad Junaid Khan
 
Unsupervised learning
Unsupervised learningUnsupervised learning
Unsupervised learningamalalhait
 
Probabilistic Reasoning
Probabilistic ReasoningProbabilistic Reasoning
Probabilistic ReasoningJunya Tanaka
 
lazy learners and other classication methods
lazy learners and other classication methodslazy learners and other classication methods
lazy learners and other classication methodsrajshreemuthiah
 
Overfitting & Underfitting
Overfitting & UnderfittingOverfitting & Underfitting
Overfitting & UnderfittingSOUMIT KAR
 
Instance Based Learning in Machine Learning
Instance Based Learning in Machine LearningInstance Based Learning in Machine Learning
Instance Based Learning in Machine LearningPavithra Thippanaik
 
State Space Representation and Search
State Space Representation and SearchState Space Representation and Search
State Space Representation and SearchHitesh Mohapatra
 

What's hot (20)

Advanced topics in artificial neural networks
Advanced topics in artificial neural networksAdvanced topics in artificial neural networks
Advanced topics in artificial neural networks
 
Bayesian learning
Bayesian learningBayesian learning
Bayesian learning
 
Dempster shafer theory
Dempster shafer theoryDempster shafer theory
Dempster shafer theory
 
Evaluating hypothesis
Evaluating  hypothesisEvaluating  hypothesis
Evaluating hypothesis
 
Target language in compiler design
Target language in compiler designTarget language in compiler design
Target language in compiler design
 
Back propagation
Back propagationBack propagation
Back propagation
 
MACHINE LEARNING-LEARNING RULE
MACHINE LEARNING-LEARNING RULEMACHINE LEARNING-LEARNING RULE
MACHINE LEARNING-LEARNING RULE
 
Learning sets of rules, Sequential Learning Algorithm,FOIL
Learning sets of rules, Sequential Learning Algorithm,FOILLearning sets of rules, Sequential Learning Algorithm,FOIL
Learning sets of rules, Sequential Learning Algorithm,FOIL
 
Issues in knowledge representation
Issues in knowledge representationIssues in knowledge representation
Issues in knowledge representation
 
Register allocation and assignment
Register allocation and assignmentRegister allocation and assignment
Register allocation and assignment
 
Unification and Lifting
Unification and LiftingUnification and Lifting
Unification and Lifting
 
Problem reduction AND OR GRAPH & AO* algorithm.ppt
Problem reduction AND OR GRAPH & AO* algorithm.pptProblem reduction AND OR GRAPH & AO* algorithm.ppt
Problem reduction AND OR GRAPH & AO* algorithm.ppt
 
Knowledge representation and Predicate logic
Knowledge representation and Predicate logicKnowledge representation and Predicate logic
Knowledge representation and Predicate logic
 
Support Vector Machines ( SVM )
Support Vector Machines ( SVM ) Support Vector Machines ( SVM )
Support Vector Machines ( SVM )
 
Unsupervised learning
Unsupervised learningUnsupervised learning
Unsupervised learning
 
Probabilistic Reasoning
Probabilistic ReasoningProbabilistic Reasoning
Probabilistic Reasoning
 
lazy learners and other classication methods
lazy learners and other classication methodslazy learners and other classication methods
lazy learners and other classication methods
 
Overfitting & Underfitting
Overfitting & UnderfittingOverfitting & Underfitting
Overfitting & Underfitting
 
Instance Based Learning in Machine Learning
Instance Based Learning in Machine LearningInstance Based Learning in Machine Learning
Instance Based Learning in Machine Learning
 
State Space Representation and Search
State Space Representation and SearchState Space Representation and Search
State Space Representation and Search
 

Similar to Using prior knowledge to initialize the hypothesis,kbann

ML_Unit_2_Part_A
ML_Unit_2_Part_AML_Unit_2_Part_A
ML_Unit_2_Part_ASrimatre K
 
Deep Belief nets
Deep Belief netsDeep Belief nets
Deep Belief netsbutest
 
Generative models
Generative modelsGenerative models
Generative modelsAvner Gidron
 
Artificial Neural Networks 1
Artificial Neural Networks 1Artificial Neural Networks 1
Artificial Neural Networks 1swapnac12
 
Instance based learning
Instance based learningInstance based learning
Instance based learningswapnac12
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfSowmyaJyothi3
 
ML_ Unit 2_Part_B
ML_ Unit 2_Part_BML_ Unit 2_Part_B
ML_ Unit 2_Part_BSrimatre K
 
Searches for Lorentz Invariance Violation with Gamma-Ray Observatories
Searches for Lorentz Invariance Violation with Gamma-Ray ObservatoriesSearches for Lorentz Invariance Violation with Gamma-Ray Observatories
Searches for Lorentz Invariance Violation with Gamma-Ray ObservatoriesVlasios Vasileiou
 
A HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEM
A HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEMA HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEM
A HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEMEM Legacy
 
Theories of error back propagation in the brain review
Theories of error back propagation in the brain reviewTheories of error back propagation in the brain review
Theories of error back propagation in the brain reviewSeonghyun Kim
 
Backpropagation.pptx
Backpropagation.pptxBackpropagation.pptx
Backpropagation.pptxVandanaVipparthi
 
rbm_hls
rbm_hlsrbm_hls
rbm_hlsDavid Lau
 
nural network ER. Abhishek k. upadhyay
nural network ER. Abhishek  k. upadhyaynural network ER. Abhishek  k. upadhyay
nural network ER. Abhishek k. upadhyayabhishek upadhyay
 
Protein Threading
Protein ThreadingProtein Threading
Protein ThreadingSANJANA PANDEY
 
Array diagnosis using compressed sensing in near field
Array diagnosis using compressed sensing in near fieldArray diagnosis using compressed sensing in near field
Array diagnosis using compressed sensing in near fieldAlexander Decker
 
Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio Armando Vieira
 
Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...
Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...
Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...IOSR Journals
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural networkNagarajan
 
B.sc biochem i bobi u 3.2 algorithm + blast
B.sc biochem i bobi u 3.2 algorithm + blastB.sc biochem i bobi u 3.2 algorithm + blast
B.sc biochem i bobi u 3.2 algorithm + blastRai University
 

Similar to Using prior knowledge to initialize the hypothesis,kbann (20)

ML_Unit_2_Part_A
ML_Unit_2_Part_AML_Unit_2_Part_A
ML_Unit_2_Part_A
 
Deep Belief nets
Deep Belief netsDeep Belief nets
Deep Belief nets
 
Generative models
Generative modelsGenerative models
Generative models
 
Artificial Neural Networks 1
Artificial Neural Networks 1Artificial Neural Networks 1
Artificial Neural Networks 1
 
Instance based learning
Instance based learningInstance based learning
Instance based learning
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
 
ML_ Unit 2_Part_B
ML_ Unit 2_Part_BML_ Unit 2_Part_B
ML_ Unit 2_Part_B
 
Searches for Lorentz Invariance Violation with Gamma-Ray Observatories
Searches for Lorentz Invariance Violation with Gamma-Ray ObservatoriesSearches for Lorentz Invariance Violation with Gamma-Ray Observatories
Searches for Lorentz Invariance Violation with Gamma-Ray Observatories
 
A HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEM
A HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEMA HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEM
A HYBRID GENETIC ALGORITHM APPROACH FOR OSPF WEIGHT SETTING PROBLEM
 
Theories of error back propagation in the brain review
Theories of error back propagation in the brain reviewTheories of error back propagation in the brain review
Theories of error back propagation in the brain review
 
Backpropagation.pptx
Backpropagation.pptxBackpropagation.pptx
Backpropagation.pptx
 
rbm_hls
rbm_hlsrbm_hls
rbm_hls
 
nural network ER. Abhishek k. upadhyay
nural network ER. Abhishek  k. upadhyaynural network ER. Abhishek  k. upadhyay
nural network ER. Abhishek k. upadhyay
 
Unit iii update
Unit iii updateUnit iii update
Unit iii update
 
Protein Threading
Protein ThreadingProtein Threading
Protein Threading
 
Array diagnosis using compressed sensing in near field
Array diagnosis using compressed sensing in near fieldArray diagnosis using compressed sensing in near field
Array diagnosis using compressed sensing in near field
 
Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio Hidden Layer Leraning Vector Quantizatio
Hidden Layer Leraning Vector Quantizatio
 
Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...
Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...
Using Multi-layered Feed-forward Neural Network (MLFNN) Architecture as Bidir...
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
B.sc biochem i bobi u 3.2 algorithm + blast
B.sc biochem i bobi u 3.2 algorithm + blastB.sc biochem i bobi u 3.2 algorithm + blast
B.sc biochem i bobi u 3.2 algorithm + blast
 

More from swapnac12

Awt, Swing, Layout managers
Awt, Swing, Layout managersAwt, Swing, Layout managers
Awt, Swing, Layout managersswapnac12
 
Applet
 Applet Applet
Appletswapnac12
 
Event handling
Event handlingEvent handling
Event handlingswapnac12
 
Asymptotic notations(Big O, Omega, Theta )
Asymptotic notations(Big O, Omega, Theta )Asymptotic notations(Big O, Omega, Theta )
Asymptotic notations(Big O, Omega, Theta )swapnac12
 
Performance analysis(Time & Space Complexity)
Performance analysis(Time & Space Complexity)Performance analysis(Time & Space Complexity)
Performance analysis(Time & Space Complexity)swapnac12
 
Introduction ,characteristics, properties,pseudo code conventions
Introduction ,characteristics, properties,pseudo code conventionsIntroduction ,characteristics, properties,pseudo code conventions
Introduction ,characteristics, properties,pseudo code conventionsswapnac12
 
Genetic algorithms
Genetic algorithmsGenetic algorithms
Genetic algorithmsswapnac12
 
Computational learning theory
Computational learning theoryComputational learning theory
Computational learning theoryswapnac12
 
Introdution and designing a learning system
Introdution and designing a learning systemIntrodution and designing a learning system
Introdution and designing a learning systemswapnac12
 
Inductive bias
Inductive biasInductive bias
Inductive biasswapnac12
 
Concept learning and candidate elimination algorithm
Concept learning and candidate elimination algorithmConcept learning and candidate elimination algorithm
Concept learning and candidate elimination algorithmswapnac12
 

More from swapnac12 (11)

Awt, Swing, Layout managers
Awt, Swing, Layout managersAwt, Swing, Layout managers
Awt, Swing, Layout managers
 
Applet
 Applet Applet
Applet
 
Event handling
Event handlingEvent handling
Event handling
 
Asymptotic notations(Big O, Omega, Theta )
Asymptotic notations(Big O, Omega, Theta )Asymptotic notations(Big O, Omega, Theta )
Asymptotic notations(Big O, Omega, Theta )
 
Performance analysis(Time & Space Complexity)
Performance analysis(Time & Space Complexity)Performance analysis(Time & Space Complexity)
Performance analysis(Time & Space Complexity)
 
Introduction ,characteristics, properties,pseudo code conventions
Introduction ,characteristics, properties,pseudo code conventionsIntroduction ,characteristics, properties,pseudo code conventions
Introduction ,characteristics, properties,pseudo code conventions
 
Genetic algorithms
Genetic algorithmsGenetic algorithms
Genetic algorithms
 
Computational learning theory
Computational learning theoryComputational learning theory
Computational learning theory
 
Introdution and designing a learning system
Introdution and designing a learning systemIntrodution and designing a learning system
Introdution and designing a learning system
 
Inductive bias
Inductive biasInductive bias
Inductive bias
 
Concept learning and candidate elimination algorithm
Concept learning and candidate elimination algorithmConcept learning and candidate elimination algorithm
Concept learning and candidate elimination algorithm
 

Recently uploaded

POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
internship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerinternship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerunnathinaik
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 
Crayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon ACrayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon AUnboundStockton
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxmanuelaromero2013
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxAvyJaneVismanos
 
Meghan Sutherland In Media Res Media Component
Meghan Sutherland In Media Res Media ComponentMeghan Sutherland In Media Res Media Component
Meghan Sutherland In Media Res Media ComponentInMediaRes1
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxEyham Joco
 
ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)Dr. Mazin Mohamed alkathiri
 
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdfFraming an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdfUjwalaBharambe
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatYousafMalik24
 
Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceSamikshaHamane
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxGaneshChakor2
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...M56BOOKSTORE PRODUCT/SERVICE
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Celine George
 

Recently uploaded (20)

POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
internship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developerinternship ppt on smartinternz platform as salesforce developer
internship ppt on smartinternz platform as salesforce developer
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 
Crayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon ACrayon Activity Handout For the Crayon A
Crayon Activity Handout For the Crayon A
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
How to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptxHow to Make a Pirate ship Primary Education.pptx
How to Make a Pirate ship Primary Education.pptx
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptx
 
Meghan Sutherland In Media Res Media Component
Meghan Sutherland In Media Res Media ComponentMeghan Sutherland In Media Res Media Component
Meghan Sutherland In Media Res Media Component
 
Types of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptxTypes of Journalistic Writing Grade 8.pptx
Types of Journalistic Writing Grade 8.pptx
 
ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)ESSENTIAL of (CS/IT/IS) class 06 (database)
ESSENTIAL of (CS/IT/IS) class 06 (database)
 
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdfFraming an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice great
 
Roles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in PharmacovigilanceRoles & Responsibilities in Pharmacovigilance
Roles & Responsibilities in Pharmacovigilance
 
Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
CARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptxCARE OF CHILD IN INCUBATOR..........pptx
CARE OF CHILD IN INCUBATOR..........pptx
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
KSHARA STURA .pptx---KSHARA KARMA THERAPY (CAUSTIC THERAPY)————IMP.OF KSHARA ...
 
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
Incoming and Outgoing Shipments in 1 STEP Using Odoo 17
 

Using prior knowledge to initialize the hypothesis,kbann

  • 1. USING PRIOR KNOWLEDGE TO INITIALIZE THE HYPOTHESIS Swapna,.C
  • 2. 3 USING PRIOR KNOWLEDGE TO INITIALIZE THE HYPOTHESIS • One approach to using prior knowledge is to initialize the hypothesis to perfectly fit the domain theory, then inductively refine this initial hypothesis as needed to fit the training data. This approach is used by the KBANN (Knowledge-Based Artificial Neural Network) algorithm to learn artificial neural networks. • In KBANN an initial network is first constructed so that for every possible instance, the classification assigned by the network is identical to that assigned by the domain theory. The BACKPROPAGATION Algorithm is then employed to adjust the weights of this initial network as needed to fit the training examples.
  • 3. • In KBANN , if the domain theory is only approximately correct, initializing the network to fit this domain theory will give a better starting approximation to the target function than initializing the network to random initial weights. This should lead, in turn, to better generalization accuracy for the final hypothesis. • Initialize-the-hypothesis approach to using the domain theory has been explored by several researchers, including Shavlik and Towel1 (1989), Towel1 and Shavlik (1994), Fu (1989, 1993), and Pratt (1993a, 1993b). We will use the KBANN algorithm described in Shavlik and Towel1 (1989) to illustrate this approach.
  • 4. 3.1 The KBANN Algorithm • The KBANN algorithm exemplifies the initialize-the-hypothesis approach to using domain theories. It assumes a domain theory represented by a set of propositional, nonrecursive Horn clauses. A Horn clause is propositional if it contains no variables.
  • 5. • Given: • A set of training examples • A domain theory consisting of nonrecursive, propositional Horn clauses • Determine: • An artificial neural network that fits the training examples, biased by the domain theory. • The two stages of the KBANN algorithm are first to create an artificial neural network that perfectly fits the domain theory and second to use the BACKPROPACATION algorithm to refine this initial network to fit the training examples.
  • 6.
  • 7. 3.2 An Illustrative Example • Each instance describes a physical object in terms of the material from which it is made, whether it is light, etc. The task is to learn the target concept Cup defined over such physical objects. The domain theory defines a Cup as an object that is Stable, Liftable, and an OpenVessel. • The domain theory also defines each of these three attributes in terms of more primitive attributes, terminating in the primitive, operational attributes that describe the instances. The domain theory is not perfectly consistent with the training examples.
  • 8.
  • 9. • For example, the domain theory fails to classify the second and third training examples as positive examples. Nevertheless, the domain theory forms a useful approximation to the target concept. • KBANN uses the domain theory and training examples together to learn the target concept more accurately than it could from either alone. • In the first stage of the KBANN algorithm (steps 1-3 in the algorithm), an initial network is constructed that is consistent with the domain theory.
  • 10.
  • 11. •For example, the network constructed from the Cup domain theory is shown in Figure. In general the network is constructed by creating a sigmoid threshold unit for each Horn clause in the domain theory. •KBANN follows the convention that a sigmoid output value greater than 0.5 is interpreted as True and a value below 0.5 as False. Each unit is therefore constructed so that its output will be greater than 0.5 just in those cases where the corresponding Horn clause applies. For each antecedent to the Horn clause, an input is created to the corresponding sigmoid unit. The weights of the sigmoid unit are then set so that it computes the logical AND of its inputs. In particular, for each input corresponding to a non-negated antecedent, the weight is set to some positive constant W. For each input corresponding to a negated antecedent, the weight is set to - W. The threshold weight of the unit, wo is then set to -(n- .5) W, where n is the number of non-negated antecedents. When unit input values are 1 or 0, this assures that their weighted sum plus wo will be positive (and the sigmoid output will therefore be greater than 0.5) if and only if all clause antecedents are satisfied.
  • 12.
  • 13. •KBANN algorithm can correctly encode the domain theory for arbitrarily deep networks. Towell and Shavlik (1994) report using W = 4.0 in many of their experiments. •Each sigmoid unit input is connected to the appropriate network input or to the output of another sigmoid unit, to mirror the graph of dependencies among the corresponding attributes in the domain theory. As a final step many additional inputs are added to each threshold unit, with their weights set approximately to zero. The role of these additional connections is to enable the network to inductively learn additional dependencies beyond those suggested by the given domain theory. The solid lines in the network of Figure indicate unit inputs with weights of W, whereas the lightly shaded lines indicate connections with initial weights near zero. It is easy to verify that for sufficiently large values of W this network will output values identical to the predictions of the domain theory.
  • 14.
  • 15. In the Cup example, however, the domain theory and training data are inconsistent, and this step therefore alters the initial network weights. The resulting trained network is summarized in Figure with dark solid lines indicating the largest positive weights, dashed lines indicating the largest negative weights, and light lines indicating negligible weights. Although the initial network misclassifies several training examples from Table , the refined network of Figure perfectly classifies all of these training examples. In Figure significant new dependencies were discovered during the inductive step, including the dependency of the Liftable unit on the feature MadeOfStyrofoam. It is important to keep in mind that while the unit labeled Liftable was initially defined by the given Horn clause for Liftable, the subsequent weight changes performed by BACKPROPAGATION have dramatically changed, the meaning of this hidden unit. After training of the network, this unit may take on a very different meaning unrelated to the initial notion of Liftable.
  • 16. Remarks: •KBANN analytically creates a network equivalent to the given domain theory, then inductively refines this initial hypothesis to better fit the training data. In doing so, it modifies the network weights as needed to overcome inconsistencies between the domain theory and observed data. •The chief benefit of KBANN over purely inductive BACKPROPAGATION beginning with random initial weights) is that it typically generalizes more accurately than BACKPROPAGATION when given an approximately correct domain theory, especially when training data is scarce..
  • 17. For example, Towel1 et al. (1990) describe the application of KBANN to a molecular genetics problem. Here the task was to learn to recognize DNA segments called promoter regions, which influence gene activity. In this experiment KBANN was given an initial domain theory obtained from a molecular geneticist, and a set of 53 positive and 53 negative training examples of promoter regions. Performance was evaluated using a leave-one-out strategy in which the system was run 106 different times. On each iteration KBANN was trained using 105 of the 106 examples and tested on the remaining example. The results of these 106 experiments were accumulated to provide an estimate of the true error rate. KBANN obtained an error rate of 41106, compared to an error rate of 81106 using standard BACKPROPAGATAION variant of the KBANN approach was applied by Fu (1993), who reports an error rate of 21106 on the same data. Thus, the impact of prior knowledge in these experiments was to reduce significantly the error rate. The training data for this experiment is available at World WideWeb site http:llwww.ics.uci.edu/~mlearn/MLRepository.htm
  • 18. To understand the significance of KBANN it is useful to consider how its hypothesis search differs from that of the purely inductive BACKPROPAGATION algorithm. The hypothesis space search conducted by both algorithms is depicted schematically in Figure
  • 19. The key difference is the initial hypothesis from which weight tuning is performed. In the case that multiple hypotheses (weight vectors) can be found that fit the data-a condition that will be especially likely when training data is scarce-KBANN is likely to converge to a hypothesis that generalizes beyond the data in a way that is more similar to the domain theory predictions. On the other hand, the particular hypothesis to which BACKPROPAGATION converges will more likely be a hypothesis with small weights, corresponding roughly to a generalization bias of smoothly interpolating between training examples. In brief, KBANN uses a domain-specific theory to bias generalization, whereas BACKPROPAGATION uses a domain-independent syntactic bias toward small weight values. Note in this summary we have ignored the effect of local minima on the search.
  • 20. Limitations of KBANN include the fact that it can accommodate only propositional domain theories; that is, collections of variable-free Horn clauses. It is also possible for KBANN to be misled when given highly inaccurate domain theories, so that its generalization accuracy can deteriorate below the level of BACKPROPAGATION. KBANN illustrates the initialize-the-hypothesis approach to combining analytical and inductive learning. These approaches vary in the exact technique for constructing the initial network, the application of BACKPROPAGATION to weight tuning, and in methods for extracting symbolic descriptions from the refined network. Pratt (1993a, 1993b) describes an initialize-the-hypothesis approach in which the prior knowledge is provided by a previously learned neural network for a related task, rather than a manually provided symbolic domain theory. Here the prior knowledge corresponds to a set of conditional independence assumptions that determine the graph structure of the Bayes net, whose conditional probability tables are then induced from the training data.