SlideShare a Scribd company logo
An Introduction to
Artificial Neural Network
Dr Iman Ardekani
Biological Neurons
Modeling Neurons
McCulloch-Pitts Neuron
Network Architecture
Learning Process
Perceptron
Linear Neuron
Multilayer Perceptron
Content
Ramón-y-Cajal (Spanish Scientist, 1852~1934):
1. Brain is composed of individual cells called
neurons.
2. Neurons are connected to each others by
synopses.
Biological Neurons
Neurons Structure (Biology)
Biological Neurons
Dendrites Cell body
Nucleus
Axon
Axon
Terminals
Synaptic Junction (Biology)
Biological Neurons
Synapse
Neurons Function (Biology)
1. Dendrites receive signal from other neurons.
2. Neurons can process (or transfer) the received signals.
3. Axon terminals deliverer the processed signal to other
tissues.
Biological Neurons
What kind of signals? Electrical Impulse Signals
Modeling Neurons (Computer Science)
Biological Neurons
Input Outputs
Process
Modeling Neurons
Net input signal is a linear combination of input signals xi.
Each Output is a function of the net input signal.
Biological Neurons
x1x2
x3
x4
x5
x6
x7
x8
x9
x10
y
- McCulloch and Pitts (1943) for introducing the
idea of neural networks as computing machines
- Hebb (1949) for inventing the first rule for self-
organized learning
- Rosenblass (1958) for proposing the perceptron
as the first model for learning with a teacher
Modelling Neurons
Net input signal received through synaptic junctions is
net = b + Σwixi = b + WT
X
Weight vector: W =[w1 w2 … wm]T
Input vector: X = [x1 x2 … xm]T
Each output is a function of the net stimulus signal (f is called the
activation function)
y = f (net) = f(b + WTX)
Modelling Neurons
General Model for Neurons
Modelling Neurons
y= f(b+ Σwnxn)
x1
x2
xm
.
.
. y
w1
w2
wm
b
 f
Activation functions
Modelling Neurons
1 1 1
1 1 1
-1 -1 -1
Good for classification Simple computation Continuous & Differentiable
Threshold Function/ Hard Limiter Linear Function sigmoid Function
Sigmoid Function
Modelling Neurons
1
sigmoid Function
f(x) =
1
1 + e-ax
= threshold function
when a goes to infinity
McCulloch-Pitts Neuron
Modelling Neurons
y {0,1}
y
x1
x2
xm
.
.
. y
w1
w2
wm
b

Modelling Neurons
y [0,1]

x1
x2
xm
.
.
. y
w1
w2
wm
b
Single-input McCulloch-Pitts neurode with b=0, w1=-1 for binary
inputs:
Conclusion?
McCulloch-Pitts Neuron
x1 net y
0 0 1
1 -1 0 x1 y
w1
b

Two-input McCulloch-Pitts neurode with b=-1,
w1=w2=1 for binary inputs:
McCulloch-Pitts Neuron
x1 x2 net y
0 0 ? ?
0 1 ? ?
1 0 ? ?
1 1 ? ?
x1
y
w1
b

x2
w2
Two-input McCulloch-Pitts neurode with b=-1,
w1=w2=1 for binary inputs:
McCulloch-Pitts Neuron
x1 x2 net y
0 0 -1 0
0 1 0 1
1 0 0 1
1 1 1 1
x1
y
w1
b

x2
w2
Two-input McCulloch-Pitts neurode with b=-2,
w1=w2=1 for binary inputs :
McCulloch-Pitts Neuron
x1 x2 net y
0 0 ? ?
0 1 ? ?
1 0 ? ?
1 1 ? ?
x1
y
w1
b

x2
w2
Two-input McCulloch-Pitts neurode with b=-2,
w1=w2=1 for binary inputs :
McCulloch-Pitts Neuron
x1 x2 net y
0 0 -2 0
0 1 -1 0
1 0 -1 0
1 1 0 1
x1
y
w1
b

x2
w2
Every basic Boolean function can be
implemented using combinations of McCulloch-
Pitts Neurons.
McCulloch-Pitts Neuron
the McCulloch-Pitts neuron can be used as a
classifier that separate the input signals into two
classes (perceptron):
McCulloch-Pitts Neuron
X1
X2
Class A (y=1)
Class B (y=0)
Class A  y=?
y = 1  net = ?
net  0  ?
b+w1x1+w2x2  0
Class B  y=?
y = 0  net = ?
net < 0  ?
b+w1x1+w2x2 < 0
Class A  b+w1x1+w2x2  0
Class B  b+w1x1+w2x2 < 0
Therefore, the decision boundary is a hyperline given by
b+w1x1+w2x2 = 0
Where w1 and w2 come from?
McCulloch-Pitts Neuron
Solution: More Neurons Required
McCulloch-Pitts Neuron
X2
X1
X1 and x2
X2
X1
X1 or x2
X2
X1
X1 xor x2
?
Nonlinear Classification
McCulloch-Pitts Neuron
X1
X2
Class A (y=1)
Class B (y=0)
Single Layer Feed-forward Network
Network Architecture
Single Layer:
There is only one
computational
layer.
Feed-forward:
Input layer projects
to the output layer
not vice versa.
 f
 f
 f
Input layer
(sources)
Output layer
(Neurons)
Multi Layer Feed-forward Network
Network Architecture
 f
 f
 f
Input layer
(sources)
Hidden layer
(Neurons)
 f
 f
Output layer
(Neurons)
Single Layer Recurrent Network
Network Architecture
 f
 f
 f
Multi Layer Recurrent Network
Network Architecture
 f
 f
 f
 f
 f
The mechanism based on which a neural network
can adjust its weights (synaptic junctions weights):
- Supervised learning: having a teacher
- Unsupervised learning: without teacher
Learning Processes
Supervised Learning
Learning Processes
Teacher
Learner
Environment 
Input
data Desired
response
Actual
response
Error Signal
+
-
Unsupervised Learning
Learning Processes
LearnerEnvironment
Input
data
Neurons learn based on a competitive task.
A competition rule is required (competitive-learning rule).
- The goal of the perceptron to classify input data into two classes A and B
- Only when the two classes can be separated by a linear boundary
- The perceptron is built around the McCulloch-Pitts Neuron model
- A linear combiner followed by a hard limiter
- Accordingly the neuron can produce +1 and 0
Perceptron
x1
x2
xm
.
.
. y
wk
wk
wk
b

Equivalent Presentation
Perceptron
x1
x2
xm
.
.
. y
w1
w2
wm
b

w0
1
net = WTX
Weight vector: W =[w0 w1 … wm]T
Input vector: X = [1 x1 x2 … xm]T
There exist a weight vector w such that we may state
WTx > 0 for every input vector x belonging to A
WTx ≤ 0 for every input vector x belonging to B
Perceptron
Elementary perceptron Learning Algorithm
1) Initiation: w(0) = a random weight vector
2) At time index n, form the input vector x(n)
3) IF (wTx > 0 and x belongs to A) or (wTx ≤ 0 and x
belongs to B) THEN w(n)=w(n-1) Otherwise
w(n)=w(n-1)-ηx(n)
4) Repeat 2 until w(n) converges
1)
Perceptron
- when the activation function simply is f(x)=x the
neuron acts similar to an adaptive filter.
- In this case: y=net=wTx
- w=[w1 w2 … wm]T
- x=[x1 x2 … xm]T
Linear Neuron
x1
x2
xm
.
.
. y
w1
w2
wm
b

Linear Neuron Learning (Adaptive Filtering)
Linear Neuron
Teacher
Learner (Linear
Neuron)
Environment 
Input
data Desired
Response
d(n)
Actual
Response
y(n)
Error Signal
+
-
LMS Algorithm
To minimize the value of the cost function defined as
E(w) = 0.5 e2(n)
where e(n) is the error signal
e(n)=d(n)-y(n)=d(n)-wT(n)x(n)
In this case, the weight vector can be updated as follows
wi(n+1)=wi(n-1) - μ( )
Linear Neuron
dE
dwi
LMS Algorithm (continued)
= e(n) = e(n) {d(n)-wT(n)x(n)}
= -e(n) xi(n)
wi(n+1)=wi(n)+μe(n)xi(n)
w(n+1)=w(n)+μe(n)x(n)
Linear Neuron
dE
dwi
de(n)
dwi
d
dwi
Summary of LMS Algorithm
1) Initiation: w(0) = a random weight vector
2) At time index n, form the input vector x(n)
3) y(n)=wT(n)x(n)
4) e(n)=d(n)-y(n)
5) w(n+1)=w(n)+μe(n)x(n)
6) Repeat 2 until w(n) converges
Linear Neuron
- To solve the xor problem
- To solve the nonlinear classification problem
- To deal with more complex problems
Multilayer Perceptron
- The activation function in multilayer perceptron is
usually a Sigmoid function.
- Because Sigmoid is differentiable function, unlike
the hard limiter function used in the elementary
perceptron.
Multilayer Perceptron
Architecture
Multilayer Perceptron
 f
 f
 f
Input layer
(sources)
Hidden layer
(Neurons)
 f
 f
Output layer
(Neurons)
Architecture
Multilayer Perceptron
 f
 f
 f
Inputs (from layer k-1) layer k Outputs (to layer k+1)
Consider a single neuron in a multilayer perceptron (neuron k)
Multilayer Perceptron
 f
yk
y0=1
y1
ym
Wk,0
Wk,1
Wk,m
.
.
.
netk = Σwk,iyi
yk= f (netk)
Multilayer perceptron Learning (Back Propagation Algorithm)
Multilayer Perceptron
Teacher
Neuron j
of layer k
layer k-1 
Input
data Desired
Response
dk(n)
Actual
Response
yk(n)
Error Signal
ek(n)
+
-
Back Propagation Algorithm:
Cost function of neuron j:
Ek = 0.5 e2
j
Cost function of network:
E = ΣEj = 0.5Σe2
j
ek = dk – yk
ek = dk – f(netk)
ek = dk – f(Σwk,iyi)
Multilayer Perceptron
Back Propagation Algorithm:
Cost function of neuron k:
Ek = 0.5 e2
k
ek = dk – yk
ek = dk – f(netk)
ek = dk – f(Σwk,iyi)
Multilayer Perceptron
Cost function of network:
E = ΣEj = 0.5Σe2
j
Multilayer Perceptron
Back Propagation Algorithm:
To minimize the value of the cost function E(wk,i), the
weight vector can be updated using a gradient based
algorithm as follows
wk,i(n+1)=wk,i (n) - μ( )
=?
Multilayer Perceptron
dE
dwk,i
dE
dwk,j
Back Propagation Algorithm:
= =
= δk yk
δk = = -ekf’(netk)
Multilayer Perceptron
dE
dwk,i
dE
dek
dek
dyk
dyk
dnetk
dnetk
dwk,i
dE
dnetk
dnetk
dwk,i
dE
dek
dek
dyk
dyk
dnetk
dE
dek
= ek
dek
dnetk
= f’(netk)
dek
dyk
= -1
dnetk
dwk,i
= yi
Local Gradient
Back Propagation Algorithm:
Substituting into the gradient-based algorithm:
wk,i(n+1)=wk,i (n) – μ δk yk
δk = -ekf’(netk) = - {dk- yk} f’(netk) = ?
If k is an output neuron we have all the terms of δk
When k is a hidden neuron?
Multilayer Perceptron
dE
dwk,i
Back Propagation Algorithm:
When k is hidden
δk = =
= f’(netk)
=?
Multilayer Perceptron
dE
dek
dek
dyk
dyk
dnetk
dE
dyk
dyk
dnetk
dek
dnetk
= f’(netk)
dE
dyk
dE
dyk
= ?
E = 0.5Σe2
j
Multilayer Perceptron
dE
dyk
dE
dyk
dej
dyk
dej
dnetj
dnetj
dyk
wjk
-f’(netj)
= -Σ ej f’(netj) wjk = -Σ δj wjk
= Σ ej
= Σ ej
We had δk =-
Substituting = Σ δj wjk into δk results in
δk = - f’(netk) Σ δj wjk
which gives the local gradient for the hidden neuron k
Multilayer Perceptron
dE
dyk
f’(netk)
dE
dyk
Summary of Back Propagation Algorithm
1) Initiation: w(0) = a random weight vector
At time index n, get the input data and
2) Calculate netk and yk for all the neurons
3) For output neurons, calculate ek=dk-yk
4) For output neurons, δk = -ekf’(netk)
5) For hidden neurons, δk = - f’(netk) Σ δj wjk
6) Update every neuron weights by
wk,i(NEW)=wk,i (OLD) – μ δk yk
7. Repeat steps 2~6 for the next data set (next time index)
Multilayer Perceptron
THE END

More Related Content

What's hot

Artifical Neural Network and its applications
Artifical Neural Network and its applicationsArtifical Neural Network and its applications
Artifical Neural Network and its applications
Sangeeta Tiwari
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
Akash Goel
 
Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...
Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...
Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...
Simplilearn
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
Knoldus Inc.
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
Ismail El Gayar
 
Deep Learning - CNN and RNN
Deep Learning - CNN and RNNDeep Learning - CNN and RNN
Deep Learning - CNN and RNN
Ashray Bhandare
 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descent
Muhammad Rasel
 
What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...
What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...
What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...
Simplilearn
 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)
EdutechLearners
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
Nagarajan
 
Artificial Neural Networks - ANN
Artificial Neural Networks - ANNArtificial Neural Networks - ANN
Artificial Neural Networks - ANN
Mohamed Talaat
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
Deep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and RegularizationDeep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and Regularization
Yan Xu
 
Mc culloch pitts neuron
Mc culloch pitts neuronMc culloch pitts neuron
Autoencoders
AutoencodersAutoencoders
Autoencoders
CloudxLab
 
Introduction to Neural Networks
Introduction to Neural NetworksIntroduction to Neural Networks
Introduction to Neural Networks
Databricks
 
Artificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
Artificial Neural Networks Lect5: Multi-Layer Perceptron & BackpropagationArtificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
Artificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
Mohammed Bennamoun
 
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
Simplilearn
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
Ashray Bhandare
 
Convolutional Neural Network (CNN) - image recognition
Convolutional Neural Network (CNN)  - image recognitionConvolutional Neural Network (CNN)  - image recognition
Convolutional Neural Network (CNN) - image recognition
YUNG-KUEI CHEN
 

What's hot (20)

Artifical Neural Network and its applications
Artifical Neural Network and its applicationsArtifical Neural Network and its applications
Artifical Neural Network and its applications
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
 
Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...
Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...
Artificial Neural Network | Deep Neural Network Explained | Artificial Neural...
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Deep Learning - CNN and RNN
Deep Learning - CNN and RNNDeep Learning - CNN and RNN
Deep Learning - CNN and RNN
 
Feed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descentFeed forward ,back propagation,gradient descent
Feed forward ,back propagation,gradient descent
 
What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...
What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...
What Is A Neural Network? | How Deep Neural Networks Work | Neural Network Tu...
 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
Artificial Neural Networks - ANN
Artificial Neural Networks - ANNArtificial Neural Networks - ANN
Artificial Neural Networks - ANN
 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
 
Deep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and RegularizationDeep Feed Forward Neural Networks and Regularization
Deep Feed Forward Neural Networks and Regularization
 
Mc culloch pitts neuron
Mc culloch pitts neuronMc culloch pitts neuron
Mc culloch pitts neuron
 
Autoencoders
AutoencodersAutoencoders
Autoencoders
 
Introduction to Neural Networks
Introduction to Neural NetworksIntroduction to Neural Networks
Introduction to Neural Networks
 
Artificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
Artificial Neural Networks Lect5: Multi-Layer Perceptron & BackpropagationArtificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
Artificial Neural Networks Lect5: Multi-Layer Perceptron & Backpropagation
 
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
What Is Deep Learning? | Introduction to Deep Learning | Deep Learning Tutori...
 
Convolutional Neural Networks
Convolutional Neural NetworksConvolutional Neural Networks
Convolutional Neural Networks
 
Convolutional Neural Network (CNN) - image recognition
Convolutional Neural Network (CNN)  - image recognitionConvolutional Neural Network (CNN)  - image recognition
Convolutional Neural Network (CNN) - image recognition
 

Viewers also liked

Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
DEEPASHRI HK
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
stellajoseph
 
Neural network & its applications
Neural network & its applications Neural network & its applications
Neural network & its applications
Ahmed_hashmi
 
artificial neural network
artificial neural networkartificial neural network
artificial neural network
Pallavi Yadav
 
Artificial Intelligence: Artificial Neural Networks
Artificial Intelligence: Artificial Neural NetworksArtificial Intelligence: Artificial Neural Networks
Artificial Intelligence: Artificial Neural Networks
The Integral Worm
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
Mohammed Bennamoun
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
NikitaRuhela
 
Artificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSArtificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKS
REHMAT ULLAH
 
neural network
neural networkneural network
neural network
STUDENT
 
SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1
sravanthi computers
 
14. mohsin dalvi artificial neural networks presentation
14. mohsin dalvi   artificial neural networks presentation14. mohsin dalvi   artificial neural networks presentation
14. mohsin dalvi artificial neural networks presentation
Purnesh Aloni
 
Perceptron
PerceptronPerceptron
Perceptron
Nagarajan
 
Back propagation
Back propagationBack propagation
Back propagation
Nagarajan
 
Neural network
Neural networkNeural network
Neural network
Silicon
 
Neural networks
Neural networksNeural networks
Neural networks
Slideshare
 
Keynote Session 3: Learning to See
Keynote Session 3: Learning to SeeKeynote Session 3: Learning to See
Keynote Session 3: Learning to See
NHSScotlandEvent2013
 
mohsin dalvi artificial neural networks questions
mohsin dalvi   artificial neural networks questionsmohsin dalvi   artificial neural networks questions
mohsin dalvi artificial neural networks questions
Akash Maurya
 
neural networks
 neural networks neural networks
Neural Networks - Types of Neurons
Neural Networks - Types of NeuronsNeural Networks - Types of Neurons
Neural Networks - Types of Neurons
Christopher Sharkey
 
Two-step Classification method for Spatial Decision Tree
Two-step Classification method for Spatial Decision TreeTwo-step Classification method for Spatial Decision Tree
Two-step Classification method for Spatial Decision Tree
Abhishek Agrawal
 

Viewers also liked (20)

Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
 
Neural network & its applications
Neural network & its applications Neural network & its applications
Neural network & its applications
 
artificial neural network
artificial neural networkartificial neural network
artificial neural network
 
Artificial Intelligence: Artificial Neural Networks
Artificial Intelligence: Artificial Neural NetworksArtificial Intelligence: Artificial Neural Networks
Artificial Intelligence: Artificial Neural Networks
 
Artificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rulesArtificial Neural Networks Lect3: Neural Network Learning rules
Artificial Neural Networks Lect3: Neural Network Learning rules
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
Artificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKSArtificial intelligence NEURAL NETWORKS
Artificial intelligence NEURAL NETWORKS
 
neural network
neural networkneural network
neural network
 
SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1
 
14. mohsin dalvi artificial neural networks presentation
14. mohsin dalvi   artificial neural networks presentation14. mohsin dalvi   artificial neural networks presentation
14. mohsin dalvi artificial neural networks presentation
 
Perceptron
PerceptronPerceptron
Perceptron
 
Back propagation
Back propagationBack propagation
Back propagation
 
Neural network
Neural networkNeural network
Neural network
 
Neural networks
Neural networksNeural networks
Neural networks
 
Keynote Session 3: Learning to See
Keynote Session 3: Learning to SeeKeynote Session 3: Learning to See
Keynote Session 3: Learning to See
 
mohsin dalvi artificial neural networks questions
mohsin dalvi   artificial neural networks questionsmohsin dalvi   artificial neural networks questions
mohsin dalvi artificial neural networks questions
 
neural networks
 neural networks neural networks
neural networks
 
Neural Networks - Types of Neurons
Neural Networks - Types of NeuronsNeural Networks - Types of Neurons
Neural Networks - Types of Neurons
 
Two-step Classification method for Spatial Decision Tree
Two-step Classification method for Spatial Decision TreeTwo-step Classification method for Spatial Decision Tree
Two-step Classification method for Spatial Decision Tree
 

Similar to Artificial Neural Network

Artificial neural networks
Artificial neural networks Artificial neural networks
Artificial neural networks
Institute of Technology Telkom
 
Annintro
AnnintroAnnintro
Annintro
kaushaljha009
 
Classification using perceptron.pptx
Classification using perceptron.pptxClassification using perceptron.pptx
Classification using perceptron.pptx
someyamohsen3
 
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
hirokazutanaka
 
10-Perceptron.pdf
10-Perceptron.pdf10-Perceptron.pdf
10-Perceptron.pdf
ESTIBALYZJIMENEZCAST
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
Sagacious IT Solution
 
MLIP - Chapter 2 - Preliminaries to deep learning
MLIP - Chapter 2 - Preliminaries to deep learningMLIP - Chapter 2 - Preliminaries to deep learning
MLIP - Chapter 2 - Preliminaries to deep learning
Charles Deledalle
 
6
66
latest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxlatest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptx
MdMahfoozAlam5
 
tutorial.ppt
tutorial.ppttutorial.ppt
tutorial.ppt
Vara Prasad
 
Neural network
Neural networkNeural network
Neural network
marada0033
 
Artificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptx
Artificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptxArtificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptx
Artificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptx
MDYasin34
 
UofT_ML_lecture.pptx
UofT_ML_lecture.pptxUofT_ML_lecture.pptx
UofT_ML_lecture.pptx
abcdefghijklmn19
 
Soft Computering Technics - Unit2
Soft Computering Technics - Unit2Soft Computering Technics - Unit2
Soft Computering Technics - Unit2
sravanthi computers
 
Multilayer Backpropagation Neural Networks for Implementation of Logic Gates
Multilayer Backpropagation Neural Networks for Implementation of Logic GatesMultilayer Backpropagation Neural Networks for Implementation of Logic Gates
Multilayer Backpropagation Neural Networks for Implementation of Logic Gates
IJCSES Journal
 
071bct537 lab4
071bct537 lab4071bct537 lab4
071bct537 lab4
shailesh kandel
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Mohammed Bennamoun
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
ssuserab4f3e
 
19_Learning.ppt
19_Learning.ppt19_Learning.ppt
19_Learning.ppt
gnans Kgnanshek
 
neural networksNnf
neural networksNnfneural networksNnf
neural networksNnf
Sandilya Sridhara
 

Similar to Artificial Neural Network (20)

Artificial neural networks
Artificial neural networks Artificial neural networks
Artificial neural networks
 
Annintro
AnnintroAnnintro
Annintro
 
Classification using perceptron.pptx
Classification using perceptron.pptxClassification using perceptron.pptx
Classification using perceptron.pptx
 
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
 
10-Perceptron.pdf
10-Perceptron.pdf10-Perceptron.pdf
10-Perceptron.pdf
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
MLIP - Chapter 2 - Preliminaries to deep learning
MLIP - Chapter 2 - Preliminaries to deep learningMLIP - Chapter 2 - Preliminaries to deep learning
MLIP - Chapter 2 - Preliminaries to deep learning
 
6
66
6
 
latest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxlatest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptx
 
tutorial.ppt
tutorial.ppttutorial.ppt
tutorial.ppt
 
Neural network
Neural networkNeural network
Neural network
 
Artificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptx
Artificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptxArtificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptx
Artificial Neural Networks (ANNs) focusing on the perceptron Algorithm.pptx
 
UofT_ML_lecture.pptx
UofT_ML_lecture.pptxUofT_ML_lecture.pptx
UofT_ML_lecture.pptx
 
Soft Computering Technics - Unit2
Soft Computering Technics - Unit2Soft Computering Technics - Unit2
Soft Computering Technics - Unit2
 
Multilayer Backpropagation Neural Networks for Implementation of Logic Gates
Multilayer Backpropagation Neural Networks for Implementation of Logic GatesMultilayer Backpropagation Neural Networks for Implementation of Logic Gates
Multilayer Backpropagation Neural Networks for Implementation of Logic Gates
 
071bct537 lab4
071bct537 lab4071bct537 lab4
071bct537 lab4
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
19_Learning.ppt
19_Learning.ppt19_Learning.ppt
19_Learning.ppt
 
neural networksNnf
neural networksNnfneural networksNnf
neural networksNnf
 

More from Iman Ardekani

Introduction to Quantitative Research Methods
Introduction to Quantitative Research MethodsIntroduction to Quantitative Research Methods
Introduction to Quantitative Research Methods
Iman Ardekani
 
Introduction to Research Methods
Introduction to Research Methods Introduction to Research Methods
Introduction to Research Methods
Iman Ardekani
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
Iman Ardekani
 
Expert Systems
Expert SystemsExpert Systems
Expert Systems
Iman Ardekani
 
Genetic Agorithm
Genetic AgorithmGenetic Agorithm
Genetic Agorithm
Iman Ardekani
 
ANC Tutorial (2013)
ANC Tutorial (2013)ANC Tutorial (2013)
ANC Tutorial (2013)
Iman Ardekani
 
Remote Active Noise Control
Remote Active Noise Control Remote Active Noise Control
Remote Active Noise Control
Iman Ardekani
 
Adaptive Active Control of Sound in Smart Rooms (2014)
Adaptive Active Control of Sound in Smart Rooms (2014)Adaptive Active Control of Sound in Smart Rooms (2014)
Adaptive Active Control of Sound in Smart Rooms (2014)
Iman Ardekani
 

More from Iman Ardekani (8)

Introduction to Quantitative Research Methods
Introduction to Quantitative Research MethodsIntroduction to Quantitative Research Methods
Introduction to Quantitative Research Methods
 
Introduction to Research Methods
Introduction to Research Methods Introduction to Research Methods
Introduction to Research Methods
 
Artificial Intelligence
Artificial IntelligenceArtificial Intelligence
Artificial Intelligence
 
Expert Systems
Expert SystemsExpert Systems
Expert Systems
 
Genetic Agorithm
Genetic AgorithmGenetic Agorithm
Genetic Agorithm
 
ANC Tutorial (2013)
ANC Tutorial (2013)ANC Tutorial (2013)
ANC Tutorial (2013)
 
Remote Active Noise Control
Remote Active Noise Control Remote Active Noise Control
Remote Active Noise Control
 
Adaptive Active Control of Sound in Smart Rooms (2014)
Adaptive Active Control of Sound in Smart Rooms (2014)Adaptive Active Control of Sound in Smart Rooms (2014)
Adaptive Active Control of Sound in Smart Rooms (2014)
 

Recently uploaded

What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.
moosaasad1975
 
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxThe use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
MAGOTI ERNEST
 
Eukaryotic Transcription Presentation.pptx
Eukaryotic Transcription Presentation.pptxEukaryotic Transcription Presentation.pptx
Eukaryotic Transcription Presentation.pptx
RitabrataSarkar3
 
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
University of Maribor
 
BREEDING METHODS FOR DISEASE RESISTANCE.pptx
BREEDING METHODS FOR DISEASE RESISTANCE.pptxBREEDING METHODS FOR DISEASE RESISTANCE.pptx
BREEDING METHODS FOR DISEASE RESISTANCE.pptx
RASHMI M G
 
Medical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptxMedical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptx
terusbelajar5
 
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
Sérgio Sacani
 
Nucleic Acid-its structural and functional complexity.
Nucleic Acid-its structural and functional complexity.Nucleic Acid-its structural and functional complexity.
Nucleic Acid-its structural and functional complexity.
Nistarini College, Purulia (W.B) India
 
Topic: SICKLE CELL DISEASE IN CHILDREN-3.pdf
Topic: SICKLE CELL DISEASE IN CHILDREN-3.pdfTopic: SICKLE CELL DISEASE IN CHILDREN-3.pdf
Topic: SICKLE CELL DISEASE IN CHILDREN-3.pdf
TinyAnderson
 
Nucleophilic Addition of carbonyl compounds.pptx
Nucleophilic Addition of carbonyl  compounds.pptxNucleophilic Addition of carbonyl  compounds.pptx
Nucleophilic Addition of carbonyl compounds.pptx
SSR02
 
aziz sancar nobel prize winner: from mardin to nobel
aziz sancar nobel prize winner: from mardin to nobelaziz sancar nobel prize winner: from mardin to nobel
aziz sancar nobel prize winner: from mardin to nobel
İsa Badur
 
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...
Travis Hills MN
 
3D Hybrid PIC simulation of the plasma expansion (ISSS-14)
3D Hybrid PIC simulation of the plasma expansion (ISSS-14)3D Hybrid PIC simulation of the plasma expansion (ISSS-14)
3D Hybrid PIC simulation of the plasma expansion (ISSS-14)
David Osipyan
 
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdfUnveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Erdal Coalmaker
 
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptx
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptxANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptx
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptx
RASHMI M G
 
8.Isolation of pure cultures and preservation of cultures.pdf
8.Isolation of pure cultures and preservation of cultures.pdf8.Isolation of pure cultures and preservation of cultures.pdf
8.Isolation of pure cultures and preservation of cultures.pdf
by6843629
 
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
yqqaatn0
 
mô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốt
mô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốtmô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốt
mô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốt
HongcNguyn6
 
NuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyerNuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyer
pablovgd
 
Phenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvementPhenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvement
IshaGoswami9
 

Recently uploaded (20)

What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.What is greenhouse gasses and how many gasses are there to affect the Earth.
What is greenhouse gasses and how many gasses are there to affect the Earth.
 
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxThe use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
 
Eukaryotic Transcription Presentation.pptx
Eukaryotic Transcription Presentation.pptxEukaryotic Transcription Presentation.pptx
Eukaryotic Transcription Presentation.pptx
 
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
 
BREEDING METHODS FOR DISEASE RESISTANCE.pptx
BREEDING METHODS FOR DISEASE RESISTANCE.pptxBREEDING METHODS FOR DISEASE RESISTANCE.pptx
BREEDING METHODS FOR DISEASE RESISTANCE.pptx
 
Medical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptxMedical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptx
 
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
 
Nucleic Acid-its structural and functional complexity.
Nucleic Acid-its structural and functional complexity.Nucleic Acid-its structural and functional complexity.
Nucleic Acid-its structural and functional complexity.
 
Topic: SICKLE CELL DISEASE IN CHILDREN-3.pdf
Topic: SICKLE CELL DISEASE IN CHILDREN-3.pdfTopic: SICKLE CELL DISEASE IN CHILDREN-3.pdf
Topic: SICKLE CELL DISEASE IN CHILDREN-3.pdf
 
Nucleophilic Addition of carbonyl compounds.pptx
Nucleophilic Addition of carbonyl  compounds.pptxNucleophilic Addition of carbonyl  compounds.pptx
Nucleophilic Addition of carbonyl compounds.pptx
 
aziz sancar nobel prize winner: from mardin to nobel
aziz sancar nobel prize winner: from mardin to nobelaziz sancar nobel prize winner: from mardin to nobel
aziz sancar nobel prize winner: from mardin to nobel
 
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...
Travis Hills' Endeavors in Minnesota: Fostering Environmental and Economic Pr...
 
3D Hybrid PIC simulation of the plasma expansion (ISSS-14)
3D Hybrid PIC simulation of the plasma expansion (ISSS-14)3D Hybrid PIC simulation of the plasma expansion (ISSS-14)
3D Hybrid PIC simulation of the plasma expansion (ISSS-14)
 
Unveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdfUnveiling the Energy Potential of Marshmallow Deposits.pdf
Unveiling the Energy Potential of Marshmallow Deposits.pdf
 
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptx
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptxANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptx
ANAMOLOUS SECONDARY GROWTH IN DICOT ROOTS.pptx
 
8.Isolation of pure cultures and preservation of cultures.pdf
8.Isolation of pure cultures and preservation of cultures.pdf8.Isolation of pure cultures and preservation of cultures.pdf
8.Isolation of pure cultures and preservation of cultures.pdf
 
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
如何办理(uvic毕业证书)维多利亚大学毕业证本科学位证书原版一模一样
 
mô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốt
mô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốtmô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốt
mô tả các thí nghiệm về đánh giá tác động dòng khí hóa sau đốt
 
NuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyerNuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyer
 
Phenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvementPhenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvement
 

Artificial Neural Network

  • 1. An Introduction to Artificial Neural Network Dr Iman Ardekani
  • 2. Biological Neurons Modeling Neurons McCulloch-Pitts Neuron Network Architecture Learning Process Perceptron Linear Neuron Multilayer Perceptron Content
  • 3. Ramón-y-Cajal (Spanish Scientist, 1852~1934): 1. Brain is composed of individual cells called neurons. 2. Neurons are connected to each others by synopses. Biological Neurons
  • 4. Neurons Structure (Biology) Biological Neurons Dendrites Cell body Nucleus Axon Axon Terminals
  • 6. Neurons Function (Biology) 1. Dendrites receive signal from other neurons. 2. Neurons can process (or transfer) the received signals. 3. Axon terminals deliverer the processed signal to other tissues. Biological Neurons What kind of signals? Electrical Impulse Signals
  • 7. Modeling Neurons (Computer Science) Biological Neurons Input Outputs Process
  • 8. Modeling Neurons Net input signal is a linear combination of input signals xi. Each Output is a function of the net input signal. Biological Neurons x1x2 x3 x4 x5 x6 x7 x8 x9 x10 y
  • 9. - McCulloch and Pitts (1943) for introducing the idea of neural networks as computing machines - Hebb (1949) for inventing the first rule for self- organized learning - Rosenblass (1958) for proposing the perceptron as the first model for learning with a teacher Modelling Neurons
  • 10. Net input signal received through synaptic junctions is net = b + Σwixi = b + WT X Weight vector: W =[w1 w2 … wm]T Input vector: X = [x1 x2 … xm]T Each output is a function of the net stimulus signal (f is called the activation function) y = f (net) = f(b + WTX) Modelling Neurons
  • 11. General Model for Neurons Modelling Neurons y= f(b+ Σwnxn) x1 x2 xm . . . y w1 w2 wm b  f
  • 12. Activation functions Modelling Neurons 1 1 1 1 1 1 -1 -1 -1 Good for classification Simple computation Continuous & Differentiable Threshold Function/ Hard Limiter Linear Function sigmoid Function
  • 13. Sigmoid Function Modelling Neurons 1 sigmoid Function f(x) = 1 1 + e-ax = threshold function when a goes to infinity
  • 14. McCulloch-Pitts Neuron Modelling Neurons y {0,1} y x1 x2 xm . . . y w1 w2 wm b 
  • 16. Single-input McCulloch-Pitts neurode with b=0, w1=-1 for binary inputs: Conclusion? McCulloch-Pitts Neuron x1 net y 0 0 1 1 -1 0 x1 y w1 b 
  • 17. Two-input McCulloch-Pitts neurode with b=-1, w1=w2=1 for binary inputs: McCulloch-Pitts Neuron x1 x2 net y 0 0 ? ? 0 1 ? ? 1 0 ? ? 1 1 ? ? x1 y w1 b  x2 w2
  • 18. Two-input McCulloch-Pitts neurode with b=-1, w1=w2=1 for binary inputs: McCulloch-Pitts Neuron x1 x2 net y 0 0 -1 0 0 1 0 1 1 0 0 1 1 1 1 1 x1 y w1 b  x2 w2
  • 19. Two-input McCulloch-Pitts neurode with b=-2, w1=w2=1 for binary inputs : McCulloch-Pitts Neuron x1 x2 net y 0 0 ? ? 0 1 ? ? 1 0 ? ? 1 1 ? ? x1 y w1 b  x2 w2
  • 20. Two-input McCulloch-Pitts neurode with b=-2, w1=w2=1 for binary inputs : McCulloch-Pitts Neuron x1 x2 net y 0 0 -2 0 0 1 -1 0 1 0 -1 0 1 1 0 1 x1 y w1 b  x2 w2
  • 21. Every basic Boolean function can be implemented using combinations of McCulloch- Pitts Neurons. McCulloch-Pitts Neuron
  • 22. the McCulloch-Pitts neuron can be used as a classifier that separate the input signals into two classes (perceptron): McCulloch-Pitts Neuron X1 X2 Class A (y=1) Class B (y=0) Class A  y=? y = 1  net = ? net  0  ? b+w1x1+w2x2  0 Class B  y=? y = 0  net = ? net < 0  ? b+w1x1+w2x2 < 0
  • 23. Class A  b+w1x1+w2x2  0 Class B  b+w1x1+w2x2 < 0 Therefore, the decision boundary is a hyperline given by b+w1x1+w2x2 = 0 Where w1 and w2 come from? McCulloch-Pitts Neuron
  • 24. Solution: More Neurons Required McCulloch-Pitts Neuron X2 X1 X1 and x2 X2 X1 X1 or x2 X2 X1 X1 xor x2 ?
  • 26. Single Layer Feed-forward Network Network Architecture Single Layer: There is only one computational layer. Feed-forward: Input layer projects to the output layer not vice versa.  f  f  f Input layer (sources) Output layer (Neurons)
  • 27. Multi Layer Feed-forward Network Network Architecture  f  f  f Input layer (sources) Hidden layer (Neurons)  f  f Output layer (Neurons)
  • 28. Single Layer Recurrent Network Network Architecture  f  f  f
  • 29. Multi Layer Recurrent Network Network Architecture  f  f  f  f  f
  • 30. The mechanism based on which a neural network can adjust its weights (synaptic junctions weights): - Supervised learning: having a teacher - Unsupervised learning: without teacher Learning Processes
  • 31. Supervised Learning Learning Processes Teacher Learner Environment  Input data Desired response Actual response Error Signal + -
  • 32. Unsupervised Learning Learning Processes LearnerEnvironment Input data Neurons learn based on a competitive task. A competition rule is required (competitive-learning rule).
  • 33. - The goal of the perceptron to classify input data into two classes A and B - Only when the two classes can be separated by a linear boundary - The perceptron is built around the McCulloch-Pitts Neuron model - A linear combiner followed by a hard limiter - Accordingly the neuron can produce +1 and 0 Perceptron x1 x2 xm . . . y wk wk wk b 
  • 34. Equivalent Presentation Perceptron x1 x2 xm . . . y w1 w2 wm b  w0 1 net = WTX Weight vector: W =[w0 w1 … wm]T Input vector: X = [1 x1 x2 … xm]T
  • 35. There exist a weight vector w such that we may state WTx > 0 for every input vector x belonging to A WTx ≤ 0 for every input vector x belonging to B Perceptron
  • 36. Elementary perceptron Learning Algorithm 1) Initiation: w(0) = a random weight vector 2) At time index n, form the input vector x(n) 3) IF (wTx > 0 and x belongs to A) or (wTx ≤ 0 and x belongs to B) THEN w(n)=w(n-1) Otherwise w(n)=w(n-1)-ηx(n) 4) Repeat 2 until w(n) converges 1) Perceptron
  • 37. - when the activation function simply is f(x)=x the neuron acts similar to an adaptive filter. - In this case: y=net=wTx - w=[w1 w2 … wm]T - x=[x1 x2 … xm]T Linear Neuron x1 x2 xm . . . y w1 w2 wm b 
  • 38. Linear Neuron Learning (Adaptive Filtering) Linear Neuron Teacher Learner (Linear Neuron) Environment  Input data Desired Response d(n) Actual Response y(n) Error Signal + -
  • 39. LMS Algorithm To minimize the value of the cost function defined as E(w) = 0.5 e2(n) where e(n) is the error signal e(n)=d(n)-y(n)=d(n)-wT(n)x(n) In this case, the weight vector can be updated as follows wi(n+1)=wi(n-1) - μ( ) Linear Neuron dE dwi
  • 40. LMS Algorithm (continued) = e(n) = e(n) {d(n)-wT(n)x(n)} = -e(n) xi(n) wi(n+1)=wi(n)+μe(n)xi(n) w(n+1)=w(n)+μe(n)x(n) Linear Neuron dE dwi de(n) dwi d dwi
  • 41. Summary of LMS Algorithm 1) Initiation: w(0) = a random weight vector 2) At time index n, form the input vector x(n) 3) y(n)=wT(n)x(n) 4) e(n)=d(n)-y(n) 5) w(n+1)=w(n)+μe(n)x(n) 6) Repeat 2 until w(n) converges Linear Neuron
  • 42. - To solve the xor problem - To solve the nonlinear classification problem - To deal with more complex problems Multilayer Perceptron
  • 43. - The activation function in multilayer perceptron is usually a Sigmoid function. - Because Sigmoid is differentiable function, unlike the hard limiter function used in the elementary perceptron. Multilayer Perceptron
  • 44. Architecture Multilayer Perceptron  f  f  f Input layer (sources) Hidden layer (Neurons)  f  f Output layer (Neurons)
  • 45. Architecture Multilayer Perceptron  f  f  f Inputs (from layer k-1) layer k Outputs (to layer k+1)
  • 46. Consider a single neuron in a multilayer perceptron (neuron k) Multilayer Perceptron  f yk y0=1 y1 ym Wk,0 Wk,1 Wk,m . . . netk = Σwk,iyi yk= f (netk)
  • 47. Multilayer perceptron Learning (Back Propagation Algorithm) Multilayer Perceptron Teacher Neuron j of layer k layer k-1  Input data Desired Response dk(n) Actual Response yk(n) Error Signal ek(n) + -
  • 48. Back Propagation Algorithm: Cost function of neuron j: Ek = 0.5 e2 j Cost function of network: E = ΣEj = 0.5Σe2 j ek = dk – yk ek = dk – f(netk) ek = dk – f(Σwk,iyi) Multilayer Perceptron
  • 49. Back Propagation Algorithm: Cost function of neuron k: Ek = 0.5 e2 k ek = dk – yk ek = dk – f(netk) ek = dk – f(Σwk,iyi) Multilayer Perceptron
  • 50. Cost function of network: E = ΣEj = 0.5Σe2 j Multilayer Perceptron
  • 51. Back Propagation Algorithm: To minimize the value of the cost function E(wk,i), the weight vector can be updated using a gradient based algorithm as follows wk,i(n+1)=wk,i (n) - μ( ) =? Multilayer Perceptron dE dwk,i dE dwk,j
  • 52. Back Propagation Algorithm: = = = δk yk δk = = -ekf’(netk) Multilayer Perceptron dE dwk,i dE dek dek dyk dyk dnetk dnetk dwk,i dE dnetk dnetk dwk,i dE dek dek dyk dyk dnetk dE dek = ek dek dnetk = f’(netk) dek dyk = -1 dnetk dwk,i = yi Local Gradient
  • 53. Back Propagation Algorithm: Substituting into the gradient-based algorithm: wk,i(n+1)=wk,i (n) – μ δk yk δk = -ekf’(netk) = - {dk- yk} f’(netk) = ? If k is an output neuron we have all the terms of δk When k is a hidden neuron? Multilayer Perceptron dE dwk,i
  • 54. Back Propagation Algorithm: When k is hidden δk = = = f’(netk) =? Multilayer Perceptron dE dek dek dyk dyk dnetk dE dyk dyk dnetk dek dnetk = f’(netk) dE dyk dE dyk
  • 55. = ? E = 0.5Σe2 j Multilayer Perceptron dE dyk dE dyk dej dyk dej dnetj dnetj dyk wjk -f’(netj) = -Σ ej f’(netj) wjk = -Σ δj wjk = Σ ej = Σ ej
  • 56. We had δk =- Substituting = Σ δj wjk into δk results in δk = - f’(netk) Σ δj wjk which gives the local gradient for the hidden neuron k Multilayer Perceptron dE dyk f’(netk) dE dyk
  • 57. Summary of Back Propagation Algorithm 1) Initiation: w(0) = a random weight vector At time index n, get the input data and 2) Calculate netk and yk for all the neurons 3) For output neurons, calculate ek=dk-yk 4) For output neurons, δk = -ekf’(netk) 5) For hidden neurons, δk = - f’(netk) Σ δj wjk 6) Update every neuron weights by wk,i(NEW)=wk,i (OLD) – μ δk yk 7. Repeat steps 2~6 for the next data set (next time index) Multilayer Perceptron

Editor's Notes

  1. - Historically important