SlideShare a Scribd company logo
1 of 96
Chapter 5
Neural Network Concepts
and Paradigms
Chapter 5 Outline
•History
•Key elements and terminology
•Topologies
•Adaptation methods
•Recall dynamics
•Comparisons of neural and non-neural methods
History of Neural Networks
•How did they get their name?
•The Age of Camelot
•The Dark Age
•The Renaissance
•The Age of Neoconnectionism
•The Age of Computational Intelligence
Why the name Neural Networks?
•Important development done by biologists,
physiologists and psychologists (PDP group)
•But also important roots in analog computing
and signal processing, so engineering heritage
is significant
•Widrow: “Massively parallel adaptive filters”
•“Carbon-based” science may provide
important insights for reverse engineering
The Age of Camelot
•William James
•McCulloch and Pitts
•Donald Hebb
•Frank Rosenblatt
•Widrow and Hoff
William James (1890)
•American psychologist
•James’ Elementary Principle: “When two
elementary brain processes have been
active together or in immediate succession,
one of them, on re-occurring, tends to
propagate its excitement to the other.”
•Also wrote of the concept that a neuron’s
activity is a function of the sum of the inputs,
with past history contributing to the
interconnection weights
McCulloch and Pitts (1943)
Five physical assumptions:
* The activity of a neuron is an all or none process
* A certain fixed number of synapses must be
excited within the period of latent addition in order
to excite a neuron at any time, and this number is
independent of previous activity and position on
the neuron
* The only significant delay within the nervous
system is synaptic delay
* The activity of any inhibitory synapse absolutely
prevents excitation of the neuron at that time
* The structure of the net does not change with
time
These assumptions describe the “McCulloch-Pitts neuron
McCulloch and Pitts, cont’d.
In most neural networks, however,
•Binary PEs are seldom used
•There is no “fixed number of connections” that must be excited
•There is generally no time delay associated with a connection to a PE
•A single negative connection does not generally inhibit a PE’s
activation
•While a NN structure may not change, the weights change during
training, and connections are sometimes pruned; new CI tools may
routinely change network structure in the future
McCulloch and Pitts, cont’d.
So, why important?
* Proved that networks of their neurons could
represent any finite logical expression
* Used a massively parallel architecture
* Provided important foundation for further
development
Donald Hebb (1948)
•Postulated Hebbian weight updating: When cell A repeatedly helps
fire B, A’s efficiency in helping to fire B is increased.
•Made four major contributions:
* Recognized that information is stored in the
weight of the synapses
* Postulated a learning rate that is proportional to
the product of neuron activation values
* Assumed weights are symmetric
* Postulated a cell assembly theory: Repeated
simultaneous activation of weakly-connected cell
group results in a more strongly connected group
Frank Rosenblatt (late 1950s)
•Defined first computer implementation: the perceptron,
which could classify patterns by modifying its
connections
•Attracted attention of engineers and physicists, using
model of biological vision
•Defined information storage as being “in connections or
associations rather than topographic representations,”
and stated that recognition/behavior “makes use of
these new pathways which have been created,
automatically activating the appropriate response.”
•Defined both self-organizing and supervised learning
mechanisms, and in sequential pattern recognition.
Rosenblatt’s Perceptron
Random sets of input cells are connected to one AU.
Generally, feedback inhibits complement of AUs that excited it.
Goal: Activate correct RU for a given input pattern.
IBM 704 Computer - 1954
First mass-produced computer with core memory and floating-point arithmetic.
Widrow and Hoff
•Engineers (!) who simulated networks on computers and
implemented designs in hardware (Adaline and Madaline)
•Formulated Least Mean Squares (LMS) algorithm that minimizes
sum-squared error: for n inputs, adjust each input to remove 1/n
of the error...great speed improvement over perceptron
•Early applications in telecommunications (echo elimination) and
medicine (electrocardiogram analysis)
•LMS adapts weights even when classifier output is correct
The Dark Age
•Perceptrons
•Stephen Grossberg
•Shun-Ichi Amari
•Teuvo Kohonen
•Kunihiko Fukishima
Perceptrons
•Minsky and Papert (1969)
•Discusses two-layer perceptrons with threshold logic units
•“...our intuitive judgement [is] that the extension [to
multilayer perceptrons with hidden layers] is sterile.”
•Drove wedge between numerics and symbolics researchers
•Neural network funding dried up
Stephen Grossberg
•Work is abstract, theoretical and dense
•Best known for Adaptive Resonance Theory
•Developed “on-center off-surround” concept
•Distinguished short-term (activation) and long-term
(weights) memory; both with “forgetting”
•Introduced sigmoid response functions for cell populations
(non-negative, inhibitory)
•Computed learning from sum of desired and actual outputs
Sigmoid Function
Shun-Ichi Amari
•Developed error correction adaptation (learning) for hidden PEs
•Analysis of randomly connected networks and temporarily associative
memories
•Developed concept of neuron pools, small mutually connected groups
of neurons, as fundamental computing element
•Recent important work in competitive learning and network memory
capacity
•Most current work in blind source separation (similar work to that of
Prof. He Zhenya at Southeast University in China)
Teuvo Kohonen
•Early work similar to James Anderson
•Best known for the Self-organizing Feature Map (and its
descendants), leading unsupervised adaptation technique
•Introduced linear, continuous-valued PEs, connection weights and
inputs
•Networks with many simultaneously active input and output PEs
(image and speech processing)
•Activation patterns on multiple output PEs represent classifications
Kunihiko Fukishima
•Best known for the neocognitron, a model for a visual pattern
recognition mechanism
•Complex: multiple layers of S-cells and C-cells; effect is to
telescope input to output
•Some translation and rotation invariance
•Relatively complex structure (~10K PEs andmultiple feedforward
weight paths and feedback loops)
•Developed way to teach “hidden” PEs in limited applications
The Renaissance
•John Hopfield
•The PDP Group
* Hinton
* McClelland
* Rumelhart
* Sejnowski
* ...12 others
John Hopfield
•1982 paper helped revive the field
•From physicist’s viewpoint: synthesized,generalized, proselytized,
and pointed the way to implementation
•Developed way to minimize “energy” of network, defined stable
states
•First NNs on silicon chips built by AT&T using Hopfield net
Hopfield Networks
Can store only about 15 percent as many states as the network
has PEs
Hamming distance between two patterns should be about 50
percent of the number of PEs
Binary (1,0) values used for PE values, but Hopfield later showed
that (-1,1) can be used.
Weights are trained according to:  

patterns
jj
j
i
ij W
V
V
W )
0
(
Hopfield Networks, Continued
•The activation value of the PE is 1 if the net input is >= 0, -1
otherwise
•In 1987, Bell Labs implemented Hopfield nets on silicon chips
•Carver Mead at Cal Tech followed with VLSI nets for artificial
retinas and cochleas
Simplified 4-PE Hopfield Network
We’ll use the patterns 1 1 1 1 and –1 –1 –1 –1 as the two we’ll store.
 

patterns
j
i
ij V
V
W )
0
(Wii
The sum over the two patterns of ViVj for each weight is always 1 + 1 = 2




j
i
i
j
ij I
V
W
i
PE
input to
Net
Simplified 4-PE Hopfield Network, Cont’d.
(1,1,1,1) and (-1,-1,-1,-1) are stored patterns.
•Input (1,-1,1,1). Randomly select node to update. Always stabilizes to
(1,1,1,1).
•Input (-1,1,-1,1). Stable pattern then depends on which node is selected
first to update.
According to Robert Hecht-Nielsen:
… by the beginning of 1986, approximately one-
third of the people in the field had been brought in
directly by Hopfield or by one of his earlier
converts. …
The Parallel Distributed Processing Group
•Contributed to forming “critical mass” in neural network field
•Published three volumes; most popular books in late 1980s;
synthesized all there was to know then
•Vol. 1 Ch. 8 contains derivation of backpropagation algorithm
•Introduced Boltzmann machines
•Presented a variety of paradigms
The Age of Neoconnectionism
1987-1998
•Donald Specht developed probabilistic neural networks
•International leaders emerged, among them John Taylor (UK); Rolf
Eckmiller (Germany); Toshio Fukuda (Japan); Paul Werbos, Richard
Lippmann, Walter Freeman, Michael Arbib and Robert Marks (US); Sha
Zong, Wu Youshou, and He Zhenya (China)
•IEEE Neural Networks Council formed (now the IEEE Computational
Intelligence Society); councils also formed in several countries/regions
•Many network paradigms introduced, numerous applications developed
•Use of personal computers (PCs) proliferated
•First IEEE World Congress on Computational Intelligence in Orlando,
Florida in 1994
The Age of Computational Intelligence
1998-present
•First use of the term (in its current context) by James Bezdek in 1992
•First text on CI in 1996
•As of 1997, IEEE Transactions in all three fields: neural networks, fuzzy
systems, and evolutionary computation
•Second IEEE World Congress on CI in Anchorage in 1998
•Neural Networks Society in 2002, Computational Intelligence Society in
2006.
•Hybrid systems became common
Neural Network Theory and
Paradigms
•What are NNs and why are they useful?
•NN components and terminology
•NN topologies
•NN learning
•NN recall
•NN taxonomy
•Comparisons with other information processing methods
What NNs are and why they are useful
•In the simplest terms, NNs map input vectors to output vectors
•Operations performed include:
* Classification
* Pattern matching and completion
* Noise removal
* Optimization
* Simulation
* Control
•NNs consist of processing elements (PEs) and weighted
connections
What NNs are and why they are
useful
•Connection weights store the information
•Weight values usually determined by adaptation procedure
•Each PE acts independently of all others
•Each PE relies only on local information
•Connection pattern provides redundancy and facilitates fault
tolerance
•NNs are able to “learn” arbitrary non-linear mappings
Feedforward Neural Network
NNs are advantageous where:
•Relatively few decisions are required from a massive amount of
data
•Nonlinear mappings must be automatically acquired
•A near-optimal solution to a combinatorial optimization problem
is required quickly
Principal elements required to
specify NNs
Topology - layers and interconnections
Adaptation - how network is configured to store information
Recall - How stored information is retrieved
Neural Network Components and
Terminology
•Terminology
•Input and output patterns
•Weights
•Processing elements
•PE functions
Network Used to Illustrate Terminology
NN Components and Terminology
IEEE Computational Intelligence Society Standards Committee
Autoassociative - one pattern
Heteroassociative - two patterns
Weights provide the path and modulate the amount of signal
passed between PEs
Weights can be positive or negative
Sparse connectivity means some weights have value of zero
It is often desirable for PE to have a “bias” value
Processing Elements
•Components of NNs where computations are performed
•PEs have two important qualities:
* They require only local information
* They produce only one output value
•Output of a PE is a function of the outputs of the previous layer
and the weights from the PEs in that layer
Actually, it’s a function of a function: calculate
input to PE, then calculate PE output (activation)
Examples of PE Input Calculations
•Linear combination
•Mean variance connections
•Min-max connections
•...others
PEs with Internal and External Biases
Input Computation Types
 Linear combination connections (most common)
 Mean-variance connections
Linear Combination
 
0
j
n
i
ji
i
j W
X
f
w
x
f
y 







 

Threshold activation function
The dot product is a similarity measure:
For vectors of fixed length, maximizing dot product minimizes mean
square separation:
 
X W W X X W
j j j
    
2 2 2
2
Mean Variance

















 
 

n
i ji
i
ji
j
v
x
w
g
y
1
2
  






 

2
exp
2
x
x
g
PE Activation Functions
•Also sometimes referred to as threshold functions or
squashing functions
•Map a PEs domain to a prespecified range
•Most common functions are:
* linear function
* step function
* ramp function
* sigmoid function
* Gaussian function
•All except linear introduce a nonlinearity
Linear PE Function
x
x
f 

)
(
Step PE Function











x
if
x
if
x
f )
(


 

otherwise
0
0
1
)
(
x
if
x
f
Binary step
function
Bipolar step function replaces 0 with -1.
Ramp PE Function
















x
if
x
if
x
x
if
x
f )
(
Sigmoid PE Function
x
e
x
f 



1
1
)
(
As alpha approaches infinity, f(x) approaches a step function.
Gaussian PE Function
)
exp(
)
(
2
v
x
x
f


Neural Network Topologies
Topology: the arrangement of the PEs,
connections and patterns into a NN
NN Topologies: Terminology
Layers: All PEs in a layer have the same source(s) and
destinations for weights, and same activation function
Intralayer weights are between PEs in same layer; interlayer
weights are between PEs in different layers
Feedforward net feeds information in only one direction; if
any feedback paths exist, it is a feedback (recurrent) NN
Two-layer Networks
•n Fx PEs fully connected to p Fy PEs
•n*p weight matrix W
•Three main types:
*Feedforward pattern matchers - map input patterns to
corresponding output patterns
*Feedback pattern matchers - can accept input on either layer
*Feedforward pattern classifiers - map input patterns into one
of p classes
Two-layer Neural Networks
Multi-layer Networks
•Input layer Fx
•L hidden layers Fy1, Fy2...Fyl
•Output layer Fz
•Weights can skip over layers, but usually don’t
•Applications:
* Pattern classification
* Pattern matching
* Function approximation
* ...any nonlinear mapping is possible with
nonlinear PEs
Complex mapping can be achieved.
Multi-layer Network
Multi-layer Networks: Issues
•How many layers are sufficient?
•How many PEs needed in (each) hidden layer?
•How much data needed fro training?
Examples of Multi-layer NNs
•Backpropagation
•Neocognitron
•Probabilistic NN (radial basis function NN)
•Boltzmann machine
•Cauchy machine
•Radial basis function networks
Neural Network Adaptation
•Adaptation (def.) - Change in connection weight values
that results in information capture that can be recalled
later
Sometimes referred to as “learning”
•Supervised/unsupervised
•Offline/online
Supervised Adaptation
•Utilizes an external teacher and/or global information
* Error correction adaptation
* Hebbian adaptation
* Competitive adaptation
* Hardwired systems
•Structural and temporal adaptation both used
•Issues:
* How long to train
* How to measure results
Unsupervised Adaptation
•Also called self-organization
•No external teacher - relies on local information
•Emergent collective properties are “discovered”
•Examples:
* Self-organizing feature maps
* Competitive adaptation
Offline Adaptation
•Used by most learning techniques
•Training patterns used to condition weights prior to use;
weights then frozen
•All patterns must be present for training
On-line Adaptation
•Add information real-time, non-destructively
•Allows the network to learn in-situ
•Examples: Adaptive Resonance Theory, fuzzy min-max
Hebbian Correlations
* Based upon correlation of PE activation values
* Unbounded PE values and weights (Hebb)
(For each pattern; Ρ is learning rate)
* Bounded PE values and unbounded weights
* Uses equation above but restricts PE values
to {0,1} or {-1,1}
* Introduces nonlinearities into the system
* Examples: Hopfield and BAM
* Limited storage capacity
j
i
old
ji
new
ji y
x
w
w 


The Delta Rule (Widrow-Hoff Rule)
(A special case of Hebbian learning)
w w a
b y
ji
new
ji
old
kj ki
kj kj kj
 
 


where
Note: the value for delta is for one pattern presented to one PE
Competitive Adaptation
•Automatically creates classes for a set of input patterns
•Two-layer network
•Two-step procedure
•Couples recall process with learning process
•Fx PEs encode input pattern, each Fy PE represents a class
Researchers include:
Grossberg
Von der Malsburg
Amari
Takeuchi
Competitive Adaptation Process
1.Determine winning Fy PE using dot product (winner has
largest value).
(Input patterns and weight vectors are often normalized.
Usually, “winner takes all.”)
2. Adjust winning PE’s connection weights (reference
vector) so as to move reference vector toward input vector
 
old
ji
ki
j
old
ji
new
ji w
a
y
t
w
w 

 )
(

Example NNs: LVQ, SOFM, ART1, ART2
1
0
1












 

n
i
ji
ki
j
k
j w
a
W
A
y
Error Correction Adaptation
•Weights are adjusted in proportion to difference between target
and actual values of PE
•Two-layer net limited to linear mappings
•Multi-layer net can capture non-linear mappings
Multilayer Error Correction
Adaptation
Credit assignment problem: For how much of an output layer PE
error is each hidden layer PE responsible?
One key is to use a continuously differentiable activation function
for hidden PEs.
Multilayer Error Correction
Adaptation
•Credit assignment problem solved by using non-decreasing
and continuously differentiable activation function for PEs
•In our example, we use linear output PEs and nonlinear hidden
PEs with sigmoid activations:
output = 1/(1+exp(-input))
•We can then use the chain rule of partial differentiation
Multilayer Network
Multilayer Error Correction Adaptation,
Cont’d.
Same cost (error) function as two-layer case
 
 
 


m
k
q
j
j
k
j
k z
b
E
1 1
2
5
.
0 Summed over PEs and patterns
For linear output PE,
  






p
i
i
j
i
k
j
k
j
k
l
p
i
i
j
i
k
j
k w
y
r
where
r
f
w
y
z
1
1
,
For nonlinear hidden PE,















n
h
h
i
h
k
i
k
i
k
n
n
h
h
i
h
k
n
i
k v
a
r
where
r
f
v
a
f
y
1
1
,
)
(
inputs
Multilayer Error Correction Adaptation,
Cont’d.
 Move along error (cost) gradient to a minimum
 Output weights Fy to Fz are adjusted like the 2-layer case, since
output PEs have linear activation functions:
i
j
j
k
j
k
j
k
i
j
j
k
w
z
z
E
w
E







 
i
k
j
k
i
k
j
k
j
k
p
i
i
k
i
j
i
j
j
k
j
k
j
k
i
j
j
k
y
y
z
b
y
w
w
z
b
z
w
E

























 

)
(
2
1
1
2
Multilayer Error Correction Adaptation,
Cont’d.
 Next, adjust weights V between input and hidden layers
 Define error i
k
k
i
k r
E /

 
 where rki is the net input to the hidden PE
(consistent with the 2-layer version)
h
k
i
k
h
i
i
k
i
k
k
h
i
k
a
v
r
r
E
v
E










The key question is how to compute the values for hidden PEs
But is the derivative of the sigmoid activation function.
ki
 
 
y r f r
ki ki n ki
/  
Multilayer Error Correction Adaptation,
Cont’d.
Now use the chain rule twice:
 
























j
i
j
j
k
i
k
i
i
k
i
j
j
j
k
i
i
j
i
k
i
k
j j
k
k
j i
k
j
k
j
k
k
i
k
k
i
k
n
i
k
k
i
k
i
k
i
k
k
i
k
k
i
k
w
r
f
w
w
y
y
r
E
y
r
r
E
y
E
r
f
y
E
r
y
y
E
r
E






















)
(
therefore
so
but
Multilayer Error Correction
Adaptation, Cont’d.
But    
f r y r y y
n ki ki ki ki ki
   
 
/ 1
 


j
i
j
j
k
i
k
i
k
i
k w
y
y 
 1
So the error assigned to a hidden PE is
Hidden PE Output PE
Multilayer Error Correction
Adaptation, Cont’d.
As is the case for two-layer error correction adaptation,
)
/
(
/ 
 k i
j
j
k
i
j
j w
E
w
E 


 and 
 k j
k
j E
E
Therefore, the weight adjustment equations are:





k
i
k
j
k
old
i
j
i
j
old
i
j
new
i
j y
w
w
E
w
w 




h
k
k
i
k
old
h
i
h
i
old
h
i
new
h
i a
v
v
E
v
v 



 




  
=
where =
Back-propagation Unresolved
Issues
•Number of training parameters
•Number and configuration of hidden layer PEs
•Avoidance of local minima during training
•Relatively long training time
Extension to mean-variance
weights
•Hidden layer PE values are calculated using:
 
 
y g r r
u a
v
g x e
i i i
ih hk
ih
h
n
x
 











;
1
2
2
where , which is similar to a Gaussian function
•Output PE values are calculated using:
z y w
j i ji
i
p



1
Cluster Analysis, and Classification
Cluster analysis: Assign patterns into one of several
subgroups.
Similar patterns are within same subgroup
Differences between subgroups are maximized
Classification: Each pattern is assigned to a class. Decision
hypersurface is adjusted so as to minimize error.
Summary of Attributes
Training time – Competitive adaptation is slow; back-propagation is
very slow
Off-line/on-line – Competitive adaptation either off- or on-line; back-
propagation off-line only
Supervised/unsupervised – Back-propagation supervised;
competitive learning unsupervised
Linear/nonlinear – Back-propagation is nonlinear; competitive
learning linear only
Storage capacity (with respect to the number of weights)-
Competitive learning is fairly high; back-propagation is high
Other Information Processing
Methods
Stochastic Approximation - Same as 2- and 3-layer error
correction algorithms
Kalman Filters - Predict the next state of a system; often
used for control systems. Back-propagation is a special case
of extended Kalman filter algorithm
Linear and Nonlinear Regression - 2-layer error correction
similar to linear regression; back-propagation is similar to
non-linear regression (minimizing mean-square error is
equivalent to curve fitting)
Other Info. Processing Methods,
Cont’d.
Correlation - Compares two patterns; uses dot product; similar
to Hebbian learning
Bayes Classification - Minimize average misclassification;
probabilistic NNs and LVQ are similar
Vector Quantization - Produces a code from an n-dimensional
input pattern; LVQ is similar
Radial Basis Functions - Used in pattern classification; use
mean/variance data; implemented in radial basis function NNs
…And then there is Computational Intelligence
Preprocessing and Postprocessing
•Techniques described are useful for other computational techniques
•Selecting training, test, and validation sets
•Preparing data
•Denormalization of output data
Selecting Training, Test, and
Validation Sets
•Can split all data available randomly; can sometimes switch roles
•Training sets - Consist of typical samples and patterns; must be sufficiently
representative so hyperspace of problem is covered well, especially near
decision surfaces
•Test sets - Similar to training patterns, but not used for training; they
determine how well the NN should perform in a production environment;
should generally reflect expected probability distribution
•Validation sets - Independent from training and test sets
Preparing Data
Scaling and normalization:
•Scaling - Applying a scale factor and a offset to each raw value;
do it same for training, test, and validation sets
•Normalization - Sets the n-dimensional input vector to a total
Euclidean length of 1
•Z-axis normalization - Retains information on parameter absolute
magnitudes lost in normalization
Scaling Data
  
  Lo
A
A
Lo
Hi
A
A
A
n
mi
k
ax
km
n
mi
k
i
k
i
k 




'
Normalizing Data
 


2
'
i
k
i
k
ki
A
A
A
Normalizes to unit length, n-dimensional
Z-axis Normalization
1. Scale each parameter; [-1,1] for example
(now maximum Euclidean length for entire vector is sqrt(n))
2. Create another input dimension by creating a synthetic variable
s for each pattern
3. Calculate z-axis normalized pattern elements; now have n+1 inputs
per pattern
Note: Not good if most values are near 0, s near 1 is then the dominant
parameter in the input
Z-axis Normalization Process
1
2
n
L
s
n
A
A
k
k
i
k
i
k



 Note: Aki is scaled input
  2
/
1
1
2
 
 n
i i
k
A
L
L is the Euclidean length of the scaled input vector
Z-axis Normalization Example
4
4
1

s
Consider two 4-dimensional input patterns:
-5,5,-5,5 and -2,2,-2,2 (n=4)
Scale them: -1,1,-1,1 and -.4,.4,-.4, .4
(L=2) (L=0.8)
(Note: They would normalize to identical vectors.)
Using Z-axis normalization,
First pattern now -.5,.5,-.5,.5,0
Second pattern now -.2,.2,-.2,.2,.9165 84
.
4
64
.
1 


s
Pattern Presentation,
and Adding Noise
•Order of pattern presentation should be considered during training
•Order usually cannot be controlled in final use
•May be good to randomize training patterns, especially for online
training (can also “shuffle” patterns after each epoch)
•Adding noise can result in smoothing of target, and can be useful
when limited quantity of training data available
Postprocessing
Denormalization of output data
  
 
C
C Lo C C
Hi Lo
C
ki
ki k k
k
' max min
min

 


where Hi and Lo are max and min PE activation values
Output domain

More Related Content

Similar to CIClassCh05.ppt

Elective Neural Networks. I. The boolean brain. On a Heuristic Point of V...
Elective Neural Networks.   I. The boolean brain.   On a Heuristic Point of V...Elective Neural Networks.   I. The boolean brain.   On a Heuristic Point of V...
Elective Neural Networks. I. The boolean brain. On a Heuristic Point of V...ABINClaude
 
History of neural networks
History of neural networks History of neural networks
History of neural networks EliyasJain
 
Blue brain
Blue brainBlue brain
Blue brainRJREJESH
 
History of the Internet
History of the InternetHistory of the Internet
History of the InternetPamela Carpio
 
Neural system/ Neural Networks
Neural system/ Neural NetworksNeural system/ Neural Networks
Neural system/ Neural Networkskashmala khalil
 
Neural networks
Neural networksNeural networks
Neural networksdemotaller
 
Feed forward back propogation algorithm .pptx
Feed forward back propogation algorithm .pptxFeed forward back propogation algorithm .pptx
Feed forward back propogation algorithm .pptxneelamsanjeevkumar
 
From neural networks to deep learning
From neural networks to deep learningFrom neural networks to deep learning
From neural networks to deep learningViet-Trung TRAN
 
Fundamentals of Neural Network (Soft Computing)
Fundamentals of Neural Network (Soft Computing)Fundamentals of Neural Network (Soft Computing)
Fundamentals of Neural Network (Soft Computing)Amit Kumar Rathi
 
Intro to Neural Networks
Intro to Neural NetworksIntro to Neural Networks
Intro to Neural NetworksDean Wyatte
 
Artificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computationArtificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computationMohammed Bennamoun
 
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...Numenta
 
Artificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and FutureArtificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and FutureGrigory Sapunov
 
Training machine learning deep learning 2017
Training machine learning deep learning 2017Training machine learning deep learning 2017
Training machine learning deep learning 2017Iwan Sofana
 
[2010 10-02] intro to microprocessors[1]
[2010 10-02] intro to microprocessors[1][2010 10-02] intro to microprocessors[1]
[2010 10-02] intro to microprocessors[1]nithilgeorge
 

Similar to CIClassCh05.ppt (20)

Elective Neural Networks. I. The boolean brain. On a Heuristic Point of V...
Elective Neural Networks.   I. The boolean brain.   On a Heuristic Point of V...Elective Neural Networks.   I. The boolean brain.   On a Heuristic Point of V...
Elective Neural Networks. I. The boolean brain. On a Heuristic Point of V...
 
History of neural networks
History of neural networks History of neural networks
History of neural networks
 
Blue brain
Blue brainBlue brain
Blue brain
 
History of the Internet
History of the InternetHistory of the Internet
History of the Internet
 
Neural system/ Neural Networks
Neural system/ Neural NetworksNeural system/ Neural Networks
Neural system/ Neural Networks
 
Neural Networks: Introducton
Neural Networks: IntroductonNeural Networks: Introducton
Neural Networks: Introducton
 
Neural networks
Neural networksNeural networks
Neural networks
 
Neural networks
Neural networksNeural networks
Neural networks
 
Feed forward back propogation algorithm .pptx
Feed forward back propogation algorithm .pptxFeed forward back propogation algorithm .pptx
Feed forward back propogation algorithm .pptx
 
From neural networks to deep learning
From neural networks to deep learningFrom neural networks to deep learning
From neural networks to deep learning
 
Fundamentals of Neural Network (Soft Computing)
Fundamentals of Neural Network (Soft Computing)Fundamentals of Neural Network (Soft Computing)
Fundamentals of Neural Network (Soft Computing)
 
Intro to Neural Networks
Intro to Neural NetworksIntro to Neural Networks
Intro to Neural Networks
 
Artificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computationArtificial Neural Networks Lect1: Introduction & neural computation
Artificial Neural Networks Lect1: Introduction & neural computation
 
shenktalk.pdf
shenktalk.pdfshenktalk.pdf
shenktalk.pdf
 
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
ICMNS Presentation: Presence of high order cell assemblies in mouse visual co...
 
Neural
NeuralNeural
Neural
 
Artificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and FutureArtificial Intelligence - Past, Present and Future
Artificial Intelligence - Past, Present and Future
 
ANN.pptx
ANN.pptxANN.pptx
ANN.pptx
 
Training machine learning deep learning 2017
Training machine learning deep learning 2017Training machine learning deep learning 2017
Training machine learning deep learning 2017
 
[2010 10-02] intro to microprocessors[1]
[2010 10-02] intro to microprocessors[1][2010 10-02] intro to microprocessors[1]
[2010 10-02] intro to microprocessors[1]
 

More from Siva453615

106622.pdf planning process in management and organisational behaviour
106622.pdf planning process in management and organisational behaviour106622.pdf planning process in management and organisational behaviour
106622.pdf planning process in management and organisational behaviourSiva453615
 
Managerial economics markets unit 3 for jntuk
Managerial economics markets unit 3 for jntukManagerial economics markets unit 3 for jntuk
Managerial economics markets unit 3 for jntukSiva453615
 
Unit-1-Introduction-to-Management
Unit-1-Introduction-to-ManagementUnit-1-Introduction-to-Management
Unit-1-Introduction-to-ManagementSiva453615
 
70055256-Reward-Management.ppt
70055256-Reward-Management.ppt70055256-Reward-Management.ppt
70055256-Reward-Management.pptSiva453615
 
431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt
431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt
431239867-MOTIVATION-REWARD-SYSTEM-ppt.pptSiva453615
 
FARRoundtable.ppt
FARRoundtable.pptFARRoundtable.ppt
FARRoundtable.pptSiva453615
 
hrm ppt.pptx
hrm ppt.pptxhrm ppt.pptx
hrm ppt.pptxSiva453615
 
15_02_2023_41526209.pdf
15_02_2023_41526209.pdf15_02_2023_41526209.pdf
15_02_2023_41526209.pdfSiva453615
 
17CS008.pdf
17CS008.pdf17CS008.pdf
17CS008.pdfSiva453615
 
15_02_2023_41526209.pdf
15_02_2023_41526209.pdf15_02_2023_41526209.pdf
15_02_2023_41526209.pdfSiva453615
 
null.pptx
null.pptxnull.pptx
null.pptxSiva453615
 
HVPE 9.1 About the Course.ppt
HVPE 9.1 About the Course.pptHVPE 9.1 About the Course.ppt
HVPE 9.1 About the Course.pptSiva453615
 
UHV I Induction Program Highlights v2.pptx
UHV I Induction Program Highlights v2.pptxUHV I Induction Program Highlights v2.pptx
UHV I Induction Program Highlights v2.pptxSiva453615
 
EITK.pptx
EITK.pptxEITK.pptx
EITK.pptxSiva453615
 
SIP UHV.pptx
SIP UHV.pptxSIP UHV.pptx
SIP UHV.pptxSiva453615
 
UHV lecture 1.ppt
UHV lecture 1.pptUHV lecture 1.ppt
UHV lecture 1.pptSiva453615
 
3 Guidelines, Content _ Process of VE.ppt
3 Guidelines, Content _ Process of VE.ppt3 Guidelines, Content _ Process of VE.ppt
3 Guidelines, Content _ Process of VE.pptSiva453615
 
427_16SACAOB3_2020051805192483 (1).ppt
427_16SACAOB3_2020051805192483 (1).ppt427_16SACAOB3_2020051805192483 (1).ppt
427_16SACAOB3_2020051805192483 (1).pptSiva453615
 
Unit-3-OB-Perception-attribution-tu-ms-2018.pptx
Unit-3-OB-Perception-attribution-tu-ms-2018.pptxUnit-3-OB-Perception-attribution-tu-ms-2018.pptx
Unit-3-OB-Perception-attribution-tu-ms-2018.pptxSiva453615
 
Unit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptx
Unit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptxUnit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptx
Unit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptxSiva453615
 

More from Siva453615 (20)

106622.pdf planning process in management and organisational behaviour
106622.pdf planning process in management and organisational behaviour106622.pdf planning process in management and organisational behaviour
106622.pdf planning process in management and organisational behaviour
 
Managerial economics markets unit 3 for jntuk
Managerial economics markets unit 3 for jntukManagerial economics markets unit 3 for jntuk
Managerial economics markets unit 3 for jntuk
 
Unit-1-Introduction-to-Management
Unit-1-Introduction-to-ManagementUnit-1-Introduction-to-Management
Unit-1-Introduction-to-Management
 
70055256-Reward-Management.ppt
70055256-Reward-Management.ppt70055256-Reward-Management.ppt
70055256-Reward-Management.ppt
 
431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt
431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt
431239867-MOTIVATION-REWARD-SYSTEM-ppt.ppt
 
FARRoundtable.ppt
FARRoundtable.pptFARRoundtable.ppt
FARRoundtable.ppt
 
hrm ppt.pptx
hrm ppt.pptxhrm ppt.pptx
hrm ppt.pptx
 
15_02_2023_41526209.pdf
15_02_2023_41526209.pdf15_02_2023_41526209.pdf
15_02_2023_41526209.pdf
 
17CS008.pdf
17CS008.pdf17CS008.pdf
17CS008.pdf
 
15_02_2023_41526209.pdf
15_02_2023_41526209.pdf15_02_2023_41526209.pdf
15_02_2023_41526209.pdf
 
null.pptx
null.pptxnull.pptx
null.pptx
 
HVPE 9.1 About the Course.ppt
HVPE 9.1 About the Course.pptHVPE 9.1 About the Course.ppt
HVPE 9.1 About the Course.ppt
 
UHV I Induction Program Highlights v2.pptx
UHV I Induction Program Highlights v2.pptxUHV I Induction Program Highlights v2.pptx
UHV I Induction Program Highlights v2.pptx
 
EITK.pptx
EITK.pptxEITK.pptx
EITK.pptx
 
SIP UHV.pptx
SIP UHV.pptxSIP UHV.pptx
SIP UHV.pptx
 
UHV lecture 1.ppt
UHV lecture 1.pptUHV lecture 1.ppt
UHV lecture 1.ppt
 
3 Guidelines, Content _ Process of VE.ppt
3 Guidelines, Content _ Process of VE.ppt3 Guidelines, Content _ Process of VE.ppt
3 Guidelines, Content _ Process of VE.ppt
 
427_16SACAOB3_2020051805192483 (1).ppt
427_16SACAOB3_2020051805192483 (1).ppt427_16SACAOB3_2020051805192483 (1).ppt
427_16SACAOB3_2020051805192483 (1).ppt
 
Unit-3-OB-Perception-attribution-tu-ms-2018.pptx
Unit-3-OB-Perception-attribution-tu-ms-2018.pptxUnit-3-OB-Perception-attribution-tu-ms-2018.pptx
Unit-3-OB-Perception-attribution-tu-ms-2018.pptx
 
Unit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptx
Unit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptxUnit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptx
Unit-6-OB-Motivation-Stress-managment-tu-ms-2018.pptx
 

Recently uploaded

IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...RajaP95
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Christo Ananth
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝soniya singh
 
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)Suman Mia
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130Suhani Kapoor
 
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSHARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSRajkumarAkumalla
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...ranjana rawat
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxJoĂŁo Esperancinha
 
High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...
High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...
High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...Call girls in Ahmedabad High profile
 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
 
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...Dr.Costas Sachpazis
 
Introduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxIntroduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxupamatechverse
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSISrknatarajan
 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSKurinjimalarL3
 
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINEDJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINEslot gacor bisa pakai pulsa
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...Call Girls in Nagpur High Profile
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINESIVASHANKAR N
 
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service NashikCall Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service NashikCall Girls in Nagpur High Profile
 
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...ranjana rawat
 

Recently uploaded (20)

IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
IMPLICATIONS OF THE ABOVE HOLISTIC UNDERSTANDING OF HARMONY ON PROFESSIONAL E...
 
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
Call for Papers - African Journal of Biological Sciences, E-ISSN: 2663-2187, ...
 
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
Model Call Girl in Narela Delhi reach out to us at 🔝8264348440🔝
 
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)Software Development Life Cycle By  Team Orange (Dept. of Pharmacy)
Software Development Life Cycle By Team Orange (Dept. of Pharmacy)
 
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
VIP Call Girls Service Kondapur Hyderabad Call +91-8250192130
 
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICSHARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
HARDNESS, FRACTURE TOUGHNESS AND STRENGTH OF CERAMICS
 
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
(ANJALI) Dange Chowk Call Girls Just Call 7001035870 [ Cash on Delivery ] Pun...
 
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptxDecoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
Decoding Kotlin - Your guide to solving the mysterious in Kotlin.pptx
 
High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...
High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...
High Profile Call Girls Dahisar Arpita 9907093804 Independent Escort Service ...
 
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
(RIA) Call Girls Bhosari ( 7001035870 ) HI-Fi Pune Escorts Service
 
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
Structural Analysis and Design of Foundations: A Comprehensive Handbook for S...
 
Introduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptxIntroduction and different types of Ethernet.pptx
Introduction and different types of Ethernet.pptx
 
UNIT-III FMM. DIMENSIONAL ANALYSIS
UNIT-III FMM.        DIMENSIONAL ANALYSISUNIT-III FMM.        DIMENSIONAL ANALYSIS
UNIT-III FMM. DIMENSIONAL ANALYSIS
 
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICSAPPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
APPLICATIONS-AC/DC DRIVES-OPERATING CHARACTERISTICS
 
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
9953056974 Call Girls In South Ex, Escorts (Delhi) NCR.pdf
 
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINEDJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
DJARUM4D - SLOT GACOR ONLINE | SLOT DEMO ONLINE
 
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...Booking open Available Pune Call Girls Koregaon Park  6297143586 Call Hot Ind...
Booking open Available Pune Call Girls Koregaon Park 6297143586 Call Hot Ind...
 
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINEMANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
MANUFACTURING PROCESS-II UNIT-2 LATHE MACHINE
 
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service NashikCall Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
Call Girls Service Nashik Vaishnavi 7001305949 Independent Escort Service Nashik
 
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
The Most Attractive Pune Call Girls Budhwar Peth 8250192130 Will You Miss Thi...
 

CIClassCh05.ppt

  • 1. Chapter 5 Neural Network Concepts and Paradigms
  • 2. Chapter 5 Outline •History •Key elements and terminology •Topologies •Adaptation methods •Recall dynamics •Comparisons of neural and non-neural methods
  • 3. History of Neural Networks •How did they get their name? •The Age of Camelot •The Dark Age •The Renaissance •The Age of Neoconnectionism •The Age of Computational Intelligence
  • 4. Why the name Neural Networks? •Important development done by biologists, physiologists and psychologists (PDP group) •But also important roots in analog computing and signal processing, so engineering heritage is significant •Widrow: “Massively parallel adaptive filters” •“Carbon-based” science may provide important insights for reverse engineering
  • 5. The Age of Camelot •William James •McCulloch and Pitts •Donald Hebb •Frank Rosenblatt •Widrow and Hoff
  • 6. William James (1890) •American psychologist •James’ Elementary Principle: “When two elementary brain processes have been active together or in immediate succession, one of them, on re-occurring, tends to propagate its excitement to the other.” •Also wrote of the concept that a neuron’s activity is a function of the sum of the inputs, with past history contributing to the interconnection weights
  • 7. McCulloch and Pitts (1943) Five physical assumptions: * The activity of a neuron is an all or none process * A certain fixed number of synapses must be excited within the period of latent addition in order to excite a neuron at any time, and this number is independent of previous activity and position on the neuron * The only significant delay within the nervous system is synaptic delay * The activity of any inhibitory synapse absolutely prevents excitation of the neuron at that time * The structure of the net does not change with time These assumptions describe the “McCulloch-Pitts neuron
  • 8. McCulloch and Pitts, cont’d. In most neural networks, however, •Binary PEs are seldom used •There is no “fixed number of connections” that must be excited •There is generally no time delay associated with a connection to a PE •A single negative connection does not generally inhibit a PE’s activation •While a NN structure may not change, the weights change during training, and connections are sometimes pruned; new CI tools may routinely change network structure in the future
  • 9. McCulloch and Pitts, cont’d. So, why important? * Proved that networks of their neurons could represent any finite logical expression * Used a massively parallel architecture * Provided important foundation for further development
  • 10. Donald Hebb (1948) •Postulated Hebbian weight updating: When cell A repeatedly helps fire B, A’s efficiency in helping to fire B is increased. •Made four major contributions: * Recognized that information is stored in the weight of the synapses * Postulated a learning rate that is proportional to the product of neuron activation values * Assumed weights are symmetric * Postulated a cell assembly theory: Repeated simultaneous activation of weakly-connected cell group results in a more strongly connected group
  • 11. Frank Rosenblatt (late 1950s) •Defined first computer implementation: the perceptron, which could classify patterns by modifying its connections •Attracted attention of engineers and physicists, using model of biological vision •Defined information storage as being “in connections or associations rather than topographic representations,” and stated that recognition/behavior “makes use of these new pathways which have been created, automatically activating the appropriate response.” •Defined both self-organizing and supervised learning mechanisms, and in sequential pattern recognition.
  • 12. Rosenblatt’s Perceptron Random sets of input cells are connected to one AU. Generally, feedback inhibits complement of AUs that excited it. Goal: Activate correct RU for a given input pattern.
  • 13. IBM 704 Computer - 1954 First mass-produced computer with core memory and floating-point arithmetic.
  • 14. Widrow and Hoff •Engineers (!) who simulated networks on computers and implemented designs in hardware (Adaline and Madaline) •Formulated Least Mean Squares (LMS) algorithm that minimizes sum-squared error: for n inputs, adjust each input to remove 1/n of the error...great speed improvement over perceptron •Early applications in telecommunications (echo elimination) and medicine (electrocardiogram analysis) •LMS adapts weights even when classifier output is correct
  • 15. The Dark Age •Perceptrons •Stephen Grossberg •Shun-Ichi Amari •Teuvo Kohonen •Kunihiko Fukishima
  • 16. Perceptrons •Minsky and Papert (1969) •Discusses two-layer perceptrons with threshold logic units •“...our intuitive judgement [is] that the extension [to multilayer perceptrons with hidden layers] is sterile.” •Drove wedge between numerics and symbolics researchers •Neural network funding dried up
  • 17. Stephen Grossberg •Work is abstract, theoretical and dense •Best known for Adaptive Resonance Theory •Developed “on-center off-surround” concept •Distinguished short-term (activation) and long-term (weights) memory; both with “forgetting” •Introduced sigmoid response functions for cell populations (non-negative, inhibitory) •Computed learning from sum of desired and actual outputs
  • 19. Shun-Ichi Amari •Developed error correction adaptation (learning) for hidden PEs •Analysis of randomly connected networks and temporarily associative memories •Developed concept of neuron pools, small mutually connected groups of neurons, as fundamental computing element •Recent important work in competitive learning and network memory capacity •Most current work in blind source separation (similar work to that of Prof. He Zhenya at Southeast University in China)
  • 20. Teuvo Kohonen •Early work similar to James Anderson •Best known for the Self-organizing Feature Map (and its descendants), leading unsupervised adaptation technique •Introduced linear, continuous-valued PEs, connection weights and inputs •Networks with many simultaneously active input and output PEs (image and speech processing) •Activation patterns on multiple output PEs represent classifications
  • 21. Kunihiko Fukishima •Best known for the neocognitron, a model for a visual pattern recognition mechanism •Complex: multiple layers of S-cells and C-cells; effect is to telescope input to output •Some translation and rotation invariance •Relatively complex structure (~10K PEs andmultiple feedforward weight paths and feedback loops) •Developed way to teach “hidden” PEs in limited applications
  • 22. The Renaissance •John Hopfield •The PDP Group * Hinton * McClelland * Rumelhart * Sejnowski * ...12 others
  • 23. John Hopfield •1982 paper helped revive the field •From physicist’s viewpoint: synthesized,generalized, proselytized, and pointed the way to implementation •Developed way to minimize “energy” of network, defined stable states •First NNs on silicon chips built by AT&T using Hopfield net
  • 24. Hopfield Networks Can store only about 15 percent as many states as the network has PEs Hamming distance between two patterns should be about 50 percent of the number of PEs Binary (1,0) values used for PE values, but Hopfield later showed that (-1,1) can be used. Weights are trained according to:    patterns jj j i ij W V V W ) 0 (
  • 25. Hopfield Networks, Continued •The activation value of the PE is 1 if the net input is >= 0, -1 otherwise •In 1987, Bell Labs implemented Hopfield nets on silicon chips •Carver Mead at Cal Tech followed with VLSI nets for artificial retinas and cochleas
  • 26. Simplified 4-PE Hopfield Network We’ll use the patterns 1 1 1 1 and –1 –1 –1 –1 as the two we’ll store.    patterns j i ij V V W ) 0 (Wii The sum over the two patterns of ViVj for each weight is always 1 + 1 = 2     j i i j ij I V W i PE input to Net
  • 27. Simplified 4-PE Hopfield Network, Cont’d. (1,1,1,1) and (-1,-1,-1,-1) are stored patterns. •Input (1,-1,1,1). Randomly select node to update. Always stabilizes to (1,1,1,1). •Input (-1,1,-1,1). Stable pattern then depends on which node is selected first to update.
  • 28. According to Robert Hecht-Nielsen: … by the beginning of 1986, approximately one- third of the people in the field had been brought in directly by Hopfield or by one of his earlier converts. …
  • 29. The Parallel Distributed Processing Group •Contributed to forming “critical mass” in neural network field •Published three volumes; most popular books in late 1980s; synthesized all there was to know then •Vol. 1 Ch. 8 contains derivation of backpropagation algorithm •Introduced Boltzmann machines •Presented a variety of paradigms
  • 30. The Age of Neoconnectionism 1987-1998 •Donald Specht developed probabilistic neural networks •International leaders emerged, among them John Taylor (UK); Rolf Eckmiller (Germany); Toshio Fukuda (Japan); Paul Werbos, Richard Lippmann, Walter Freeman, Michael Arbib and Robert Marks (US); Sha Zong, Wu Youshou, and He Zhenya (China) •IEEE Neural Networks Council formed (now the IEEE Computational Intelligence Society); councils also formed in several countries/regions •Many network paradigms introduced, numerous applications developed •Use of personal computers (PCs) proliferated •First IEEE World Congress on Computational Intelligence in Orlando, Florida in 1994
  • 31. The Age of Computational Intelligence 1998-present •First use of the term (in its current context) by James Bezdek in 1992 •First text on CI in 1996 •As of 1997, IEEE Transactions in all three fields: neural networks, fuzzy systems, and evolutionary computation •Second IEEE World Congress on CI in Anchorage in 1998 •Neural Networks Society in 2002, Computational Intelligence Society in 2006. •Hybrid systems became common
  • 32. Neural Network Theory and Paradigms •What are NNs and why are they useful? •NN components and terminology •NN topologies •NN learning •NN recall •NN taxonomy •Comparisons with other information processing methods
  • 33. What NNs are and why they are useful •In the simplest terms, NNs map input vectors to output vectors •Operations performed include: * Classification * Pattern matching and completion * Noise removal * Optimization * Simulation * Control •NNs consist of processing elements (PEs) and weighted connections
  • 34. What NNs are and why they are useful •Connection weights store the information •Weight values usually determined by adaptation procedure •Each PE acts independently of all others •Each PE relies only on local information •Connection pattern provides redundancy and facilitates fault tolerance •NNs are able to “learn” arbitrary non-linear mappings
  • 36. NNs are advantageous where: •Relatively few decisions are required from a massive amount of data •Nonlinear mappings must be automatically acquired •A near-optimal solution to a combinatorial optimization problem is required quickly
  • 37. Principal elements required to specify NNs Topology - layers and interconnections Adaptation - how network is configured to store information Recall - How stored information is retrieved
  • 38. Neural Network Components and Terminology •Terminology •Input and output patterns •Weights •Processing elements •PE functions
  • 39. Network Used to Illustrate Terminology
  • 40. NN Components and Terminology IEEE Computational Intelligence Society Standards Committee Autoassociative - one pattern Heteroassociative - two patterns Weights provide the path and modulate the amount of signal passed between PEs Weights can be positive or negative Sparse connectivity means some weights have value of zero It is often desirable for PE to have a “bias” value
  • 41. Processing Elements •Components of NNs where computations are performed •PEs have two important qualities: * They require only local information * They produce only one output value •Output of a PE is a function of the outputs of the previous layer and the weights from the PEs in that layer Actually, it’s a function of a function: calculate input to PE, then calculate PE output (activation)
  • 42. Examples of PE Input Calculations •Linear combination •Mean variance connections •Min-max connections •...others
  • 43. PEs with Internal and External Biases
  • 44. Input Computation Types  Linear combination connections (most common)  Mean-variance connections
  • 45. Linear Combination   0 j n i ji i j W X f w x f y            Threshold activation function
  • 46. The dot product is a similarity measure: For vectors of fixed length, maximizing dot product minimizes mean square separation:   X W W X X W j j j      2 2 2 2
  • 47. Mean Variance                       n i ji i ji j v x w g y 1 2             2 exp 2 x x g
  • 48. PE Activation Functions •Also sometimes referred to as threshold functions or squashing functions •Map a PEs domain to a prespecified range •Most common functions are: * linear function * step function * ramp function * sigmoid function * Gaussian function •All except linear introduce a nonlinearity
  • 49. Linear PE Function x x f   ) (
  • 50. Step PE Function            x if x if x f ) (      otherwise 0 0 1 ) ( x if x f Binary step function Bipolar step function replaces 0 with -1.
  • 52. Sigmoid PE Function x e x f     1 1 ) ( As alpha approaches infinity, f(x) approaches a step function.
  • 54. Neural Network Topologies Topology: the arrangement of the PEs, connections and patterns into a NN
  • 55. NN Topologies: Terminology Layers: All PEs in a layer have the same source(s) and destinations for weights, and same activation function Intralayer weights are between PEs in same layer; interlayer weights are between PEs in different layers Feedforward net feeds information in only one direction; if any feedback paths exist, it is a feedback (recurrent) NN
  • 56. Two-layer Networks •n Fx PEs fully connected to p Fy PEs •n*p weight matrix W •Three main types: *Feedforward pattern matchers - map input patterns to corresponding output patterns *Feedback pattern matchers - can accept input on either layer *Feedforward pattern classifiers - map input patterns into one of p classes
  • 58. Multi-layer Networks •Input layer Fx •L hidden layers Fy1, Fy2...Fyl •Output layer Fz •Weights can skip over layers, but usually don’t •Applications: * Pattern classification * Pattern matching * Function approximation * ...any nonlinear mapping is possible with nonlinear PEs Complex mapping can be achieved.
  • 60. Multi-layer Networks: Issues •How many layers are sufficient? •How many PEs needed in (each) hidden layer? •How much data needed fro training?
  • 61. Examples of Multi-layer NNs •Backpropagation •Neocognitron •Probabilistic NN (radial basis function NN) •Boltzmann machine •Cauchy machine •Radial basis function networks
  • 62. Neural Network Adaptation •Adaptation (def.) - Change in connection weight values that results in information capture that can be recalled later Sometimes referred to as “learning” •Supervised/unsupervised •Offline/online
  • 63. Supervised Adaptation •Utilizes an external teacher and/or global information * Error correction adaptation * Hebbian adaptation * Competitive adaptation * Hardwired systems •Structural and temporal adaptation both used •Issues: * How long to train * How to measure results
  • 64. Unsupervised Adaptation •Also called self-organization •No external teacher - relies on local information •Emergent collective properties are “discovered” •Examples: * Self-organizing feature maps * Competitive adaptation
  • 65. Offline Adaptation •Used by most learning techniques •Training patterns used to condition weights prior to use; weights then frozen •All patterns must be present for training
  • 66. On-line Adaptation •Add information real-time, non-destructively •Allows the network to learn in-situ •Examples: Adaptive Resonance Theory, fuzzy min-max
  • 67. Hebbian Correlations * Based upon correlation of PE activation values * Unbounded PE values and weights (Hebb) (For each pattern; Ρ is learning rate) * Bounded PE values and unbounded weights * Uses equation above but restricts PE values to {0,1} or {-1,1} * Introduces nonlinearities into the system * Examples: Hopfield and BAM * Limited storage capacity j i old ji new ji y x w w   
  • 68. The Delta Rule (Widrow-Hoff Rule) (A special case of Hebbian learning) w w a b y ji new ji old kj ki kj kj kj       where Note: the value for delta is for one pattern presented to one PE
  • 69. Competitive Adaptation •Automatically creates classes for a set of input patterns •Two-layer network •Two-step procedure •Couples recall process with learning process •Fx PEs encode input pattern, each Fy PE represents a class Researchers include: Grossberg Von der Malsburg Amari Takeuchi
  • 70. Competitive Adaptation Process 1.Determine winning Fy PE using dot product (winner has largest value). (Input patterns and weight vectors are often normalized. Usually, “winner takes all.”) 2. Adjust winning PE’s connection weights (reference vector) so as to move reference vector toward input vector   old ji ki j old ji new ji w a y t w w    ) (  Example NNs: LVQ, SOFM, ART1, ART2 1 0 1                n i ji ki j k j w a W A y
  • 71. Error Correction Adaptation •Weights are adjusted in proportion to difference between target and actual values of PE •Two-layer net limited to linear mappings •Multi-layer net can capture non-linear mappings
  • 72. Multilayer Error Correction Adaptation Credit assignment problem: For how much of an output layer PE error is each hidden layer PE responsible? One key is to use a continuously differentiable activation function for hidden PEs.
  • 73. Multilayer Error Correction Adaptation •Credit assignment problem solved by using non-decreasing and continuously differentiable activation function for PEs •In our example, we use linear output PEs and nonlinear hidden PEs with sigmoid activations: output = 1/(1+exp(-input)) •We can then use the chain rule of partial differentiation
  • 75. Multilayer Error Correction Adaptation, Cont’d. Same cost (error) function as two-layer case         m k q j j k j k z b E 1 1 2 5 . 0 Summed over PEs and patterns For linear output PE,          p i i j i k j k j k l p i i j i k j k w y r where r f w y z 1 1 , For nonlinear hidden PE,                n h h i h k i k i k n n h h i h k n i k v a r where r f v a f y 1 1 , ) ( inputs
  • 76. Multilayer Error Correction Adaptation, Cont’d.  Move along error (cost) gradient to a minimum  Output weights Fy to Fz are adjusted like the 2-layer case, since output PEs have linear activation functions: i j j k j k j k i j j k w z z E w E          i k j k i k j k j k p i i k i j i j j k j k j k i j j k y y z b y w w z b z w E                             ) ( 2 1 1 2
  • 77. Multilayer Error Correction Adaptation, Cont’d.  Next, adjust weights V between input and hidden layers  Define error i k k i k r E /     where rki is the net input to the hidden PE (consistent with the 2-layer version) h k i k h i i k i k k h i k a v r r E v E           The key question is how to compute the values for hidden PEs But is the derivative of the sigmoid activation function. ki     y r f r ki ki n ki /  
  • 78. Multilayer Error Correction Adaptation, Cont’d. Now use the chain rule twice:                           j i j j k i k i i k i j j j k i i j i k i k j j k k j i k j k j k k i k k i k n i k k i k i k i k k i k k i k w r f w w y y r E y r r E y E r f y E r y y E r E                       ) ( therefore so but
  • 79. Multilayer Error Correction Adaptation, Cont’d. But     f r y r y y n ki ki ki ki ki       / 1     j i j j k i k i k i k w y y   1 So the error assigned to a hidden PE is Hidden PE Output PE
  • 80. Multilayer Error Correction Adaptation, Cont’d. As is the case for two-layer error correction adaptation, ) / ( /   k i j j k i j j w E w E     and   k j k j E E Therefore, the weight adjustment equations are:      k i k j k old i j i j old i j new i j y w w E w w      h k k i k old h i h i old h i new h i a v v E v v              = where =
  • 81. Back-propagation Unresolved Issues •Number of training parameters •Number and configuration of hidden layer PEs •Avoidance of local minima during training •Relatively long training time
  • 82. Extension to mean-variance weights •Hidden layer PE values are calculated using:     y g r r u a v g x e i i i ih hk ih h n x              ; 1 2 2 where , which is similar to a Gaussian function •Output PE values are calculated using: z y w j i ji i p    1
  • 83. Cluster Analysis, and Classification Cluster analysis: Assign patterns into one of several subgroups. Similar patterns are within same subgroup Differences between subgroups are maximized Classification: Each pattern is assigned to a class. Decision hypersurface is adjusted so as to minimize error.
  • 84. Summary of Attributes Training time – Competitive adaptation is slow; back-propagation is very slow Off-line/on-line – Competitive adaptation either off- or on-line; back- propagation off-line only Supervised/unsupervised – Back-propagation supervised; competitive learning unsupervised Linear/nonlinear – Back-propagation is nonlinear; competitive learning linear only Storage capacity (with respect to the number of weights)- Competitive learning is fairly high; back-propagation is high
  • 85. Other Information Processing Methods Stochastic Approximation - Same as 2- and 3-layer error correction algorithms Kalman Filters - Predict the next state of a system; often used for control systems. Back-propagation is a special case of extended Kalman filter algorithm Linear and Nonlinear Regression - 2-layer error correction similar to linear regression; back-propagation is similar to non-linear regression (minimizing mean-square error is equivalent to curve fitting)
  • 86. Other Info. Processing Methods, Cont’d. Correlation - Compares two patterns; uses dot product; similar to Hebbian learning Bayes Classification - Minimize average misclassification; probabilistic NNs and LVQ are similar Vector Quantization - Produces a code from an n-dimensional input pattern; LVQ is similar Radial Basis Functions - Used in pattern classification; use mean/variance data; implemented in radial basis function NNs …And then there is Computational Intelligence
  • 87. Preprocessing and Postprocessing •Techniques described are useful for other computational techniques •Selecting training, test, and validation sets •Preparing data •Denormalization of output data
  • 88. Selecting Training, Test, and Validation Sets •Can split all data available randomly; can sometimes switch roles •Training sets - Consist of typical samples and patterns; must be sufficiently representative so hyperspace of problem is covered well, especially near decision surfaces •Test sets - Similar to training patterns, but not used for training; they determine how well the NN should perform in a production environment; should generally reflect expected probability distribution •Validation sets - Independent from training and test sets
  • 89. Preparing Data Scaling and normalization: •Scaling - Applying a scale factor and a offset to each raw value; do it same for training, test, and validation sets •Normalization - Sets the n-dimensional input vector to a total Euclidean length of 1 •Z-axis normalization - Retains information on parameter absolute magnitudes lost in normalization
  • 90. Scaling Data      Lo A A Lo Hi A A A n mi k ax km n mi k i k i k      '
  • 92. Z-axis Normalization 1. Scale each parameter; [-1,1] for example (now maximum Euclidean length for entire vector is sqrt(n)) 2. Create another input dimension by creating a synthetic variable s for each pattern 3. Calculate z-axis normalized pattern elements; now have n+1 inputs per pattern Note: Not good if most values are near 0, s near 1 is then the dominant parameter in the input
  • 93. Z-axis Normalization Process 1 2 n L s n A A k k i k i k     Note: Aki is scaled input   2 / 1 1 2    n i i k A L L is the Euclidean length of the scaled input vector
  • 94. Z-axis Normalization Example 4 4 1  s Consider two 4-dimensional input patterns: -5,5,-5,5 and -2,2,-2,2 (n=4) Scale them: -1,1,-1,1 and -.4,.4,-.4, .4 (L=2) (L=0.8) (Note: They would normalize to identical vectors.) Using Z-axis normalization, First pattern now -.5,.5,-.5,.5,0 Second pattern now -.2,.2,-.2,.2,.9165 84 . 4 64 . 1    s
  • 95. Pattern Presentation, and Adding Noise •Order of pattern presentation should be considered during training •Order usually cannot be controlled in final use •May be good to randomize training patterns, especially for online training (can also “shuffle” patterns after each epoch) •Adding noise can result in smoothing of target, and can be useful when limited quantity of training data available
  • 96. Postprocessing Denormalization of output data      C C Lo C C Hi Lo C ki ki k k k ' max min min      where Hi and Lo are max and min PE activation values Output domain