SlideShare a Scribd company logo
1 of 68
Neural Networks

            Prof. Ruchi Sharma
 Bharatividyapeeth College of Engineering
                 New Delhi




                   Lecture 1                1
Plan
   Requirements
   Background
       Brains and Computers
       Computational Models
         
             Learnability vs Programming
         
             Representability vs Training Rules
   Abstractions of Neurons
   Abstractions of Networks
       Completeness of 3-Level Mc-
        Cullough-Pitts Neurons
   Learnability in Perceptrons


                      Lecture 1                   2
Brains and Computers
   What is computation?
   Basic Computational Model
    (Like C, Pascal, C++, etc)
    (See course on Models of
      Computation.)

    Analysis of problem and
     reduction to algorithm.

    Implementation of algorithm by
    Programmer.


               Lecture 1             3
Computation
Brain also computes. But
Its programming seems
   different.

Flexible, Fault Tolerant
  (neurons die every day),
Automatic Programming,
Learns from examples,
Generalizability


            Lecture 1        4
Brain vs. Computer
   Brain works on slow
    components (10**-3 sec)
   Computers use fast
    components (10**-9 sec)
   Brain more efficient (few
    joules per operation) (factor
    of 10**10.)
   Uses massively parallel
    mechanism.
   ****Can we copy its secrets?

              Lecture 1         5
Brain vs Computer
   Areas that Brain is better:
        Sense recognition and
    integration
        Working with incomplete
    information
         Generalizing
          Learning from examples
          Fault tolerant (regular
    programming is notoriously
    fragile.)


              Lecture 1        6
AI vs. NNs
   AI relates to cognitive
    psychology
       Chess, Theorem Proving,
        Expert Systems, Intelligent
        agents (e.g. on Internet)


   NNs relates to
    neurophysiology
       Recognition tasks, Associative
        Memory, Learning


                 Lecture 1            7
How can NNs work?
   Look at Brain:
   10**10 neurons (10
    Gigabytes) .
   10 ** 20 possible
    connections with different
    numbers of dendrites (reals)
   Actually about 6x 10**13
    connections (I.e. 60,000 hard
    discs for one photo of
    contents of one brain!)

              Lecture 1        8
Brain
   Complex
   Non-linear (more later!)
   Parallel Processing
   Fault Tolerant
   Adaptive
   Learns
   Generalizes
   Self Programs



              Lecture 1        9
Abstracting
   Note: Size may not be crucial
    (apylsia or crab does many
    things)
   Look at simple structures first




               Lecture 1        10
Real and Artificial
Neurons




         Lecture 1    11
One Neuron
     McCullough-Pitts

        This is very complicated. But
         abstracting the details,we
x1       have
           w1

x2        w2     Σ
          wn      Ιntegrate      Τhreshold

     xn


     Integrate-and-fire Neuron

                    Lecture 1                12
Representability

   What functions can be
    represented by a network of
    Mccullough-Pitts neurons?

   Theorem: Every logic
    function of an arbitrary
    number of variables can be
    represented by a three level
    network of neurons.



              Lecture 1        13
Proof
   Show simple functions: and,
    or, not, implies
   Recall representability of
    logic functions by DNF form.




              Lecture 1       14
AND, OR, NOT
x1              1.0
           w1
          1.0
x2        w2          Σ
          wn          Ιntegrate       Τhreshold

     xn                                   1.5




                          Lecture 1               15
AND, OR, NOT
x1              1.0
           w1
          1.0
x2        w2          Σ
          wn          Ιntegrate       Τhreshold

     xn                                   .9




                          Lecture 1               16
Lecture 1   17
AND, OR, NOT
x1             -1.0
           w1

x2        w2          Σ
          wn          Ιntegrate       Τhreshold

     xn                                   -.5




                          Lecture 1               18
DNF and All Functions
   Theorem
       Any logic (boolean) function of
        any number of variables can
        be represented in a network of
        McCullough-Pitts neurons.

       In fact the depth of the network
        is three.

       Proof: Use DNF and And, Or,
        Not representation


                 Lecture 1           19
Other Questions?

What if we allow REAL
 numbers as inputs/outputs?
What real functions can be
 represented?

What if we modify threshold to
 some other function; so
 output is not {0,1}. What
 functions can be
 represented?

            Lecture 1       20
Representability and
Generalizability




        Lecture 1      21
Learnability and
Generalizability

   The previous theorem tells
    us that neural networks are
    potentially powerful, but
    doesn’t tell us how to use
    them.
   We desire simple networks
    with uniform training rules.




              Lecture 1        22
One Neuron
(Perceptron)
   What can be represented by
    one neuron?
   Is there an automatic way to
    learn a function by
    examples?




              Lecture 1       23
Perceptron Training
Rule
 Loop:
      Take an example. Apply
  to network. If correct
  answer,
  return to loop. If incorrect,
  go to FIX.
FIX: Adjust network weights
  by
  input example . Go to Loop.


            Lecture 1        24
Example of
Perceptron
Learning
   X1 = 1 (+) x2 = -.5 (-)
   X3 = 3 (+) x4 = -2 (-)
   Expanded Vector
       Y1 = (1,1) (+) y2= (-.5,1)(-)
       Y3 = (3,1) (+) y4 = (-2,1) (-)


Random initial weight
   (-2.5, 1.75)



                  Lecture 1              25
Graph of Learning




        Lecture 1   26
Trace of Perceptron
       W1 y1 = (-2.5,1.75) (1,1)<0
        wrong
   W2 = w1 + y1 = (-1.5, 2.75)
       W2 y2 = (-1.5, 2.75)(-.5, 1)>0
        wrong
   W3 = w2 – y2 = (-1, 1.75)
       W3 y3 = (-1,1.75)(3,1) <0
        wrong
   W4 = w4 + y3 = (2, 2.75)



                 Lecture 1            27
Perceptron
Convergence
Theorem
   If the concept is
    representable in a
    perceptron then the
    perceptron learning rule will
    converge in a finite amount
    of time.

   (MAKE PRECISE and Prove)




               Lecture 1        28
What is a Neural
Network?
   What is an abstract Neuron?
   What is a Neural Network?
   How are they computed?
   What are the advantages?
   Where can they be used?

       Agenda
       What to expect




                Lecture 1    29
Perceptron Algorithm
   Start: Choose arbitrary value
    for weights, W
   Test: Choose arbitrary
    example X
   If X pos and WX >0 or X neg
    and WX <= 0 go to Test
   Fix:
       If X pos W := W +X;
       If X negative W:= W –X;
       Go to Test;



                 Lecture 1        30
Perceptron Conv.
Thm.
   Let F be a set of unit length
    vectors. If there is a vector
    V* and a value e>0 such that
     V*X > e for all X in F then
    the perceptron program goes
    to FIX only a finite number of
    times.




              Lecture 1         31
Applying Algorithm to
“And”
   W0 = (0,0,1) or random
   X1 = (0,0,1) result 0
   X2 = (0,1,1) result 0
   X3 = (1,0, 1) result 0
   X4 = (1,1,1) result 1




             Lecture 1       32
“And” continued
    Wo X1 > 0 wrong;
    W1 = W0 – X1 = (0,0,0)
    W1 X2 = 0 OK (Bdry)
    W1 X3 = 0 OK
    W1 X4 = 0 wrong;
    W2 = W1 +X4 = (1,1,1)
    W3 X1 = 1 wrong
    W4 = W3 –X1 = (1,1,0)
    W4X2 = 1 wrong
    W5 = W4 – X2 = (1,0, -1)
    W5 X3 = 0 OK
    W5 X4 = 0 wrong
    W6 = W5 + X4 = (2, 1, 0)
    W6 X1 = 0 OK
    W6 X2 = 1 wrong
    W7 = W7 – X2 = (2,0, -1)
                Lecture 1       33
“And” page 3
   W8 X3 = 1 wrong
   W9 = W8 – X3 = (1,0, 0)
   W9X4 = 1 OK
   W9 X1 = 0 OK
   W9 X2 = 0 OK
   W9 X3 = 1 wrong
   W10 = W9 – X3 = (0,0,-1)
   W10X4 = -1 wrong
   W11 = W10 + X4 = (1,1,0)
   W11X1 =0 OK
   W11X2 = 1 wrong
   W12 = W12 – X2 = (1,0, -1)

               Lecture 1         34
Proof of Conv
    Theorem
  Note:
   1. By hypothesis, there is a δ >0
such that V*X > δ for all x ε F
  1. Can eliminate threshold
 (add additional dimension to input)
   W(x,y,z) > threshold if and only if
   W* (x,y,z,1) > 0
   2. Can assume all examples are
   positive ones
  (Replace negative examples
   by their negated vectors)
   W(x,y,z) <0 if and only if
     W(-x,-y,-z) > 0.
                 Lecture 1               35
Proof (cont).
 Consider quotient V*W/|W|.
(note: this is multidimensional
cosine between V* and W.)
Recall V* is unit vector .

Quotient <= 1.




              Lecture 1           36
Proof(cont)
Now each time FIX is visited W
 changes via ADD.
V* W(n+1) = V*(W(n) + X)
          = V* W(n) + V*X
          >= V* W(n) + δ
Hence

V* W(n) >= n δ       (∗)




             Lecture 1            37
Proof (cont)
   Now consider denominator:
   |W(n+1)| = W(n+1)W(n+1) =
    ( W(n) + X)(W(n) + X) = |W(n)|
    **2 + 2W(n)X + 1

    (recall |X| = 1 and W(n)X < 0
    since X is positive example and
    we are in FIX)
     < |W(n)|**2 + 1

So after n times
  |W(n+1)|**2 < n           (**)


                Lecture 1          38
Proof (cont)
   Putting (*) and (**) together:

    Quotient = V*W/|W|
          > nδ/ sqrt(n)

Since Quotient <=1 this means
   n < (1/δ)∗∗2.
This means we enter FIX a
  bounded number of times.
              Q.E.D.


                 Lecture 1           39
Geometric Proof
   See hand slides.




              Lecture 1   40
Perceptron Diagram 1


               Solution
                 OK



No examples               No examples


               No examples




              Lecture 1           41
Solution
                 OK



No examples               No examples


               No examples




              Lecture 1           42
Lecture 1   43
Additional Facts
   Note: If X’s presented in
    systematic way, then solution W
    always found.
   Note: Not necessarily same as V*
   Note: If F not finite, may not
    obtain solution in finite time
   Can modify algorithm in minor
    ways and stays valid (e.g. not
    unit but bounded examples);
    changes in W(n).



               Lecture 1         44
Perceptron
Convergence
Theorem
   If the concept is
    representable in a
    perceptron then the
    perceptron learning rule will
    converge in a finite amount
    of time.

   (MAKE PRECISE and Prove)




               Lecture 1        45
Important Points
   Theorem only guarantees
    result IF representable!
   Usually we are not interested
    in just representing
    something but in its
    generalizability – how will it
    work on examples we havent
    seen!




              Lecture 1         46
Percentage of Boolean
Functions Representible by
a Perceptron


   Input   Perceptron Functions
      1         4           4
      2         16          14
      3         104         256
      4       1,882       65,536
      5       94,572         10**9
      6    15,028,134        10**19
      7 8,378,070,864        10**38
      8 17,561,539,552,946    10**77




              Lecture 1           47
Generalizability
   Typically train a network on a
    sample set of examples
   Use it on general class
   Training can be slow; but
    execution is fast.




              Lecture 1        48
Perceptron
    •weights




   ∑w x i i   > threshold ⇔ A in receptive field
                      kdkdkfjlll
 ∑w x
    i    i    > θ ⇔ The letter A is in the recept

         •Pattern
         Identification
         •(Note: Neuron
         is trained)


                        Lecture 1                  49
What wont work?
   Example: Connectedness
    with bounded diameter
    perceptron.

 Compare with Convex with
(use sensors of order three).




             Lecture 1          50
Biases and
Thresholds
   We can
    replace the
    threshold with
    a bias.                       n

                                 ∑ wx >θ    i i
   A bias acts                  i= 1
                                  n
    exactly as a
    weight on a
                                 ∑ w x -θ > 0
                                 i= 1
                                            i i
                           n
    connection
    from a unit          ∑ wx > 0
                          i= 0
                                      i i         w0 = - θ x0 = 1
    whose
    activation is
    always 1.


                    Lecture 1                                  51
Perceptron
   Loop :
       Take an example and apply to
        network.
       If correct answer – return to
        Loop.
       If incorrect – go to Fix.
   Fix :
       Adjust network weights by
        input example.
       Go to Loop.


                 Lecture 1          52
Perceptron Algorithm
 v     x∈ F
Let       be arbitrary
Choose: chooseF
Test: v ⋅Ifx > 0x∈  x ∈ and
                        F+                 go to
      v⋅ x ≤
   Choose 0         x∈ F +
        If               and              go to
      v⋅ x ≤ 0
   Fix plus         x∈ F   −

      vIf > 0
         ⋅x         x ∈ and
                        F−                go to
       v := v + x
   Choose
        If               and              go to
       v := v − x
   Fix minus
Fix plus:                       go to Choose
Fix minus:                      go to Choose



                    Lecture 1                  53
Perceptron Algorithm
   Conditions to the algorithm
    existence :
        Condition no.1: +
                 
    

                      if x ∈ F v * ⋅ x > 0
              ∃ v∗ , 
                      if x ∈ F v * ⋅ x ≤ 0
                                −



       Condition no.2:
        We choose F to be a group of unit
        vectors.




                   Lecture 1                   54
Geometric viewpoint


             wn + 1
         x               v   *

    wn




             Lecture 1           55
Perceptron Algorithm
   Based on these conditions the
    number of times we enter the
    Loop is finite.
                 Examples
   Proof:          world

                            +     −
                    F = F F
          { x A x > 0} { y A y ≤ 0}
              *                    *


          Positiv               Negati
            e                      ve
          exampl                examp
            es                    les




                    Lecture 1            56
Perceptron Algorithm-
Proof
   We replace the threshold
    with a bias.
   We assume F is a group
    of unit vectors.
        ∀φ ∈ F     φ =1




                 Lecture 1     57
Perceptron Algorithm-
Proof
    We reduce
                      ∑ ay≤0                    ∑ a (− y ) ≥ 0


    what we have
                            *
                            i i       ⇒             *
                                                    i   i

    to prove by                   yi ⇒ -yi
    eliminating all                         +
    the negative                     F= F
    examples and
    placing their           ∀φ ∈ F        A*φ > 0
    negations in
    the positive            ∀φ ∈ F        A*φ > δ
    examples.                        ⇓
                            ∃δ > 0        Aφ > δ
                                            *




                Lecture 1                                   58
Perceptron Algorithm-
 Proof
                            A* ⋅ A
                                A* ⋅ A
                  G ( A) = * =         ≤1
                           A ⋅A A

The
numerator :


      A ⋅ Al + 1 = A ⋅ ( Al + φ ) = A ⋅ Al + A ⋅ φ ≥ A ⋅ Al + δ
       *           *                 *       *        *


                                                 >δ
       After n
       changes                A* ⋅ An ≥ n ⋅ δ




                           Lecture 1                              59
Perceptron Algorithm-
 Proof
The
denominator
:


         At + 1 = ( At + 1 )( At + 1 ) = ( At + φ )( At + φ ) =
              2


                         2                       2       2
                  = At + 2 ⋅ At ⋅ φ + φ < At + 1
                                 <0           =1
        After n                              2
        changes                         An < n




                             Lecture 1                            60
Perceptron Algorithm-
 Proof
From the
denominato
r:                             2
                             An < n
From the
numerator :

                         A* ⋅ An > n ⋅ δ


                         A ⋅ An n ⋅ δ
                         *
                G ( A) =       >      = n ⋅δ ≤ 1
                          An       n
                                  ⇓
              n is                     1
              final
                             n<
                                   δ   2




                         Lecture 1                 61
Example - AND
                                               x1 x2   bias AND


w0 = 0,0,0
                                                      0
w0 ⋅ x 0 = w0 ⋅ 0,0,1 = 0
w0 ⋅ x1 = w0 ⋅ 0,1,1 = 0                        0 0 1 0
w0 ⋅ x 2 = w0 ⋅ 1,0,1 = 0
                             wrong             0 1 1 0
w0 ⋅ x3 = w0 ⋅ 1,1,1 = 0
FIX = w1 = w0 + x3 = 0,0,0 + 1,1,1 = 1,1,1      1 0 1 1
       w1 ⋅ x 0 = w1 ⋅ 0,0,1 = 1
FIX = w2 = w1 − x 0 = 1,1,1 − 0,0,1wrong
                               = 1,1,0
                                                1 1 1
       w2 ⋅ x1 = w2 ⋅ 0,1,1 = 1
FIX = w3 = w2 − x1 = 1,1,0 − 0,1,1 = 1,0,−1


                                    wrong



                              …etc


                                   Lecture 1                62
AND – Bi Polar
-
    solution
    -
        1   0                            -   -
1   1                                            1   0
                                         1   1
    -                                        -
1       1   0                            1       1   0
    1                                        1
-                                        -
    1   1   0                                1   1   0
1                                        1
1   1   1   0  wrong+                   1   1   1   1
-   -                            1 1 1
        1   0
1   1
    -
1       1   1
    1            wrong -                        success
-                                0 2 0
    1   1   1
1                wrong -
                                 1 1 -
1   1   1   1
                                     1


            continue

                            Lecture 1                63
Problem



                 LHS1 MHS1 RHS1 + MHS2 MHS 2RHS < LHSLHS 3 + 3MHS 33 + RHS 3 > 0
                       LHS 2       LHS + RHS 2    0 3 MHS RHS
                                      2        2

   LHS1 + MHS1 + RHS1 > 0
                        MHS1 = MHS 2 = MHS 3

LHS1 + RHS1 > 0             LHS 2 + RHS 2 < 0                  LHS 3 + RHS3 > 0

               LHS1 = LHS 2                   RHS 2 = RHS 3


       should be small
     RHS 2                                    should be small
                                            LHS 2
            enough so that                       enough so that
            LHS 2 + RHS 2 < 0                     LHS 2 + RHS 2 < 0


                               contradictio
                                       !!!n

                                   Lecture 1                                       64
Linear Separation
   Every perceptron determines
    a classification of vector
    inputs which is determined
    by a hyperline
   Two dimensional examples
    (add algebra)


    OR      AND           XOR



                           not
                         possible

             Lecture 1              65
Linear Separation in
Higher Dimensions
In higher dimensions, still
  linear separation, but hard to
  tell

Example: Connected;
 Convex - which can be
 handled by Perceptron with
 local sensors; which can not
 be.

Note: Define local sensors.

             Lecture 1        66
What wont work?
   Try XOR.




               Lecture 1   67
Limitations of
Perceptron
   Representability
       Only concepts that are linearly
        separable.
       Compare: Convex versus
        connected
       Examples: XOR vs OR




                 Lecture 1           68

More Related Content

What's hot

SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1
sravanthi computers
 
Multi-Layer Perceptrons
Multi-Layer PerceptronsMulti-Layer Perceptrons
Multi-Layer Perceptrons
ESCOM
 
Hebbian Learning
Hebbian LearningHebbian Learning
Hebbian Learning
ESCOM
 

What's hot (20)

Lecture 9 Perceptron
Lecture 9 PerceptronLecture 9 Perceptron
Lecture 9 Perceptron
 
03 Single layer Perception Classifier
03 Single layer Perception Classifier03 Single layer Perception Classifier
03 Single layer Perception Classifier
 
Sparse autoencoder
Sparse autoencoderSparse autoencoder
Sparse autoencoder
 
JAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
JAISTサマースクール2016「脳を知るための理論」講義03 Network DynamicsJAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
JAISTサマースクール2016「脳を知るための理論」講義03 Network Dynamics
 
Lesson 38
Lesson 38Lesson 38
Lesson 38
 
SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1
 
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNSArtificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
Artificial Neural Networks Lect2: Neurobiology & Architectures of ANNS
 
Perceptron in ANN
Perceptron in ANNPerceptron in ANN
Perceptron in ANN
 
Lec10
Lec10Lec10
Lec10
 
Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)
 
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
 
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rulesJAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
JAISTサマースクール2016「脳を知るための理論」講義02 Synaptic Learning rules
 
Nural network ER.Abhishek k. upadhyay
Nural network  ER.Abhishek k. upadhyayNural network  ER.Abhishek k. upadhyay
Nural network ER.Abhishek k. upadhyay
 
Unit 1
Unit 1Unit 1
Unit 1
 
Multi-Layer Perceptrons
Multi-Layer PerceptronsMulti-Layer Perceptrons
Multi-Layer Perceptrons
 
Neural network
Neural networkNeural network
Neural network
 
Introduction to Neural networks (under graduate course) Lecture 2 of 9
Introduction to Neural networks (under graduate course) Lecture 2 of 9Introduction to Neural networks (under graduate course) Lecture 2 of 9
Introduction to Neural networks (under graduate course) Lecture 2 of 9
 
neural networksNnf
neural networksNnfneural networksNnf
neural networksNnf
 
Hebbian Learning
Hebbian LearningHebbian Learning
Hebbian Learning
 
Neural network
Neural networkNeural network
Neural network
 

Similar to Nn3

latest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxlatest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptx
MdMahfoozAlam5
 
cs4811-ch11-neural-networks.ppt
cs4811-ch11-neural-networks.pptcs4811-ch11-neural-networks.ppt
cs4811-ch11-neural-networks.ppt
butest
 
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
Dongseo University
 
05 history of cv a machine learning (theory) perspective on computer vision
05  history of cv a machine learning (theory) perspective on computer vision05  history of cv a machine learning (theory) perspective on computer vision
05 history of cv a machine learning (theory) perspective on computer vision
zukun
 
Artificial Neural Network
Artificial Neural Network Artificial Neural Network
Artificial Neural Network
Iman Ardekani
 
Dm part03 neural-networks-handout
Dm part03 neural-networks-handoutDm part03 neural-networks-handout
Dm part03 neural-networks-handout
okeee
 

Similar to Nn3 (20)

ann-ics320Part4.ppt
ann-ics320Part4.pptann-ics320Part4.ppt
ann-ics320Part4.ppt
 
ann-ics320Part4.ppt
ann-ics320Part4.pptann-ics320Part4.ppt
ann-ics320Part4.ppt
 
2-Perceptrons.pdf
2-Perceptrons.pdf2-Perceptrons.pdf
2-Perceptrons.pdf
 
10-Perceptron.pdf
10-Perceptron.pdf10-Perceptron.pdf
10-Perceptron.pdf
 
latest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptxlatest TYPES OF NEURAL NETWORKS (2).pptx
latest TYPES OF NEURAL NETWORKS (2).pptx
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
 
Two algorithms to accelerate training of back-propagation neural networks
Two algorithms to accelerate training of back-propagation neural networksTwo algorithms to accelerate training of back-propagation neural networks
Two algorithms to accelerate training of back-propagation neural networks
 
Ann ics320 part4
Ann ics320 part4Ann ics320 part4
Ann ics320 part4
 
Soft Computering Technics - Unit2
Soft Computering Technics - Unit2Soft Computering Technics - Unit2
Soft Computering Technics - Unit2
 
cs4811-ch11-neural-networks.ppt
cs4811-ch11-neural-networks.pptcs4811-ch11-neural-networks.ppt
cs4811-ch11-neural-networks.ppt
 
8_Neural Networks in artificial intelligence.ppt
8_Neural Networks in artificial intelligence.ppt8_Neural Networks in artificial intelligence.ppt
8_Neural Networks in artificial intelligence.ppt
 
Perceptron
PerceptronPerceptron
Perceptron
 
neural networks
 neural networks neural networks
neural networks
 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
 
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
2013-1 Machine Learning Lecture 04 - Michael Negnevitsky - Artificial neur…
 
05 history of cv a machine learning (theory) perspective on computer vision
05  history of cv a machine learning (theory) perspective on computer vision05  history of cv a machine learning (theory) perspective on computer vision
05 history of cv a machine learning (theory) perspective on computer vision
 
071bct537 lab4
071bct537 lab4071bct537 lab4
071bct537 lab4
 
Artificial Neural Network
Artificial Neural Network Artificial Neural Network
Artificial Neural Network
 
Max net
Max netMax net
Max net
 
Dm part03 neural-networks-handout
Dm part03 neural-networks-handoutDm part03 neural-networks-handout
Dm part03 neural-networks-handout
 

More from Ruchi Sharma (6)

Rtos
RtosRtos
Rtos
 
Ipc feb4
Ipc feb4Ipc feb4
Ipc feb4
 
Ipc ppt
Ipc pptIpc ppt
Ipc ppt
 
EMBEDDED SYSTEMS
EMBEDDED SYSTEMSEMBEDDED SYSTEMS
EMBEDDED SYSTEMS
 
Nn ppt
Nn pptNn ppt
Nn ppt
 
neural networks
neural networksneural networks
neural networks
 

Recently uploaded

Salient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsSalient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functions
KarakKing
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
ZurliaSoop
 

Recently uploaded (20)

80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
 
Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structure
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptxOn_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
On_Translating_a_Tamil_Poem_by_A_K_Ramanujan.pptx
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Towards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptxTowards a code of practice for AI in AT.pptx
Towards a code of practice for AI in AT.pptx
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
Salient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsSalient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functions
 
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptxHMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
HMCS Max Bernays Pre-Deployment Brief (May 2024).pptx
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptx
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Jamworks pilot and AI at Jisc (20/03/2024)
Jamworks pilot and AI at Jisc (20/03/2024)Jamworks pilot and AI at Jisc (20/03/2024)
Jamworks pilot and AI at Jisc (20/03/2024)
 

Nn3

  • 1. Neural Networks Prof. Ruchi Sharma Bharatividyapeeth College of Engineering New Delhi Lecture 1 1
  • 2. Plan  Requirements  Background  Brains and Computers  Computational Models  Learnability vs Programming  Representability vs Training Rules  Abstractions of Neurons  Abstractions of Networks  Completeness of 3-Level Mc- Cullough-Pitts Neurons  Learnability in Perceptrons Lecture 1 2
  • 3. Brains and Computers  What is computation?  Basic Computational Model (Like C, Pascal, C++, etc) (See course on Models of Computation.) Analysis of problem and reduction to algorithm. Implementation of algorithm by Programmer. Lecture 1 3
  • 4. Computation Brain also computes. But Its programming seems different. Flexible, Fault Tolerant (neurons die every day), Automatic Programming, Learns from examples, Generalizability Lecture 1 4
  • 5. Brain vs. Computer  Brain works on slow components (10**-3 sec)  Computers use fast components (10**-9 sec)  Brain more efficient (few joules per operation) (factor of 10**10.)  Uses massively parallel mechanism.  ****Can we copy its secrets? Lecture 1 5
  • 6. Brain vs Computer  Areas that Brain is better: Sense recognition and integration Working with incomplete information Generalizing Learning from examples Fault tolerant (regular programming is notoriously fragile.) Lecture 1 6
  • 7. AI vs. NNs  AI relates to cognitive psychology  Chess, Theorem Proving, Expert Systems, Intelligent agents (e.g. on Internet)  NNs relates to neurophysiology  Recognition tasks, Associative Memory, Learning Lecture 1 7
  • 8. How can NNs work?  Look at Brain:  10**10 neurons (10 Gigabytes) .  10 ** 20 possible connections with different numbers of dendrites (reals)  Actually about 6x 10**13 connections (I.e. 60,000 hard discs for one photo of contents of one brain!) Lecture 1 8
  • 9. Brain  Complex  Non-linear (more later!)  Parallel Processing  Fault Tolerant  Adaptive  Learns  Generalizes  Self Programs Lecture 1 9
  • 10. Abstracting  Note: Size may not be crucial (apylsia or crab does many things)  Look at simple structures first Lecture 1 10
  • 12. One Neuron McCullough-Pitts  This is very complicated. But abstracting the details,we x1 have w1 x2 w2 Σ wn Ιntegrate Τhreshold xn Integrate-and-fire Neuron Lecture 1 12
  • 13. Representability  What functions can be represented by a network of Mccullough-Pitts neurons?  Theorem: Every logic function of an arbitrary number of variables can be represented by a three level network of neurons. Lecture 1 13
  • 14. Proof  Show simple functions: and, or, not, implies  Recall representability of logic functions by DNF form. Lecture 1 14
  • 15. AND, OR, NOT x1 1.0 w1 1.0 x2 w2 Σ wn Ιntegrate Τhreshold xn 1.5 Lecture 1 15
  • 16. AND, OR, NOT x1 1.0 w1 1.0 x2 w2 Σ wn Ιntegrate Τhreshold xn .9 Lecture 1 16
  • 17. Lecture 1 17
  • 18. AND, OR, NOT x1 -1.0 w1 x2 w2 Σ wn Ιntegrate Τhreshold xn -.5 Lecture 1 18
  • 19. DNF and All Functions  Theorem  Any logic (boolean) function of any number of variables can be represented in a network of McCullough-Pitts neurons.  In fact the depth of the network is three.  Proof: Use DNF and And, Or, Not representation Lecture 1 19
  • 20. Other Questions? What if we allow REAL numbers as inputs/outputs? What real functions can be represented? What if we modify threshold to some other function; so output is not {0,1}. What functions can be represented? Lecture 1 20
  • 22. Learnability and Generalizability  The previous theorem tells us that neural networks are potentially powerful, but doesn’t tell us how to use them.  We desire simple networks with uniform training rules. Lecture 1 22
  • 23. One Neuron (Perceptron)  What can be represented by one neuron?  Is there an automatic way to learn a function by examples? Lecture 1 23
  • 24. Perceptron Training Rule  Loop: Take an example. Apply to network. If correct answer, return to loop. If incorrect, go to FIX. FIX: Adjust network weights by input example . Go to Loop. Lecture 1 24
  • 25. Example of Perceptron Learning  X1 = 1 (+) x2 = -.5 (-)  X3 = 3 (+) x4 = -2 (-)  Expanded Vector  Y1 = (1,1) (+) y2= (-.5,1)(-)  Y3 = (3,1) (+) y4 = (-2,1) (-) Random initial weight (-2.5, 1.75) Lecture 1 25
  • 26. Graph of Learning Lecture 1 26
  • 27. Trace of Perceptron  W1 y1 = (-2.5,1.75) (1,1)<0 wrong  W2 = w1 + y1 = (-1.5, 2.75)  W2 y2 = (-1.5, 2.75)(-.5, 1)>0 wrong  W3 = w2 – y2 = (-1, 1.75)  W3 y3 = (-1,1.75)(3,1) <0 wrong  W4 = w4 + y3 = (2, 2.75) Lecture 1 27
  • 28. Perceptron Convergence Theorem  If the concept is representable in a perceptron then the perceptron learning rule will converge in a finite amount of time.  (MAKE PRECISE and Prove) Lecture 1 28
  • 29. What is a Neural Network?  What is an abstract Neuron?  What is a Neural Network?  How are they computed?  What are the advantages?  Where can they be used?  Agenda  What to expect Lecture 1 29
  • 30. Perceptron Algorithm  Start: Choose arbitrary value for weights, W  Test: Choose arbitrary example X  If X pos and WX >0 or X neg and WX <= 0 go to Test  Fix:  If X pos W := W +X;  If X negative W:= W –X;  Go to Test; Lecture 1 30
  • 31. Perceptron Conv. Thm.  Let F be a set of unit length vectors. If there is a vector V* and a value e>0 such that V*X > e for all X in F then the perceptron program goes to FIX only a finite number of times. Lecture 1 31
  • 32. Applying Algorithm to “And”  W0 = (0,0,1) or random  X1 = (0,0,1) result 0  X2 = (0,1,1) result 0  X3 = (1,0, 1) result 0  X4 = (1,1,1) result 1 Lecture 1 32
  • 33. “And” continued  Wo X1 > 0 wrong;  W1 = W0 – X1 = (0,0,0)  W1 X2 = 0 OK (Bdry)  W1 X3 = 0 OK  W1 X4 = 0 wrong;  W2 = W1 +X4 = (1,1,1)  W3 X1 = 1 wrong  W4 = W3 –X1 = (1,1,0)  W4X2 = 1 wrong  W5 = W4 – X2 = (1,0, -1)  W5 X3 = 0 OK  W5 X4 = 0 wrong  W6 = W5 + X4 = (2, 1, 0)  W6 X1 = 0 OK  W6 X2 = 1 wrong  W7 = W7 – X2 = (2,0, -1) Lecture 1 33
  • 34. “And” page 3  W8 X3 = 1 wrong  W9 = W8 – X3 = (1,0, 0)  W9X4 = 1 OK  W9 X1 = 0 OK  W9 X2 = 0 OK  W9 X3 = 1 wrong  W10 = W9 – X3 = (0,0,-1)  W10X4 = -1 wrong  W11 = W10 + X4 = (1,1,0)  W11X1 =0 OK  W11X2 = 1 wrong  W12 = W12 – X2 = (1,0, -1) Lecture 1 34
  • 35. Proof of Conv Theorem  Note: 1. By hypothesis, there is a δ >0 such that V*X > δ for all x ε F 1. Can eliminate threshold (add additional dimension to input) W(x,y,z) > threshold if and only if W* (x,y,z,1) > 0 2. Can assume all examples are positive ones (Replace negative examples by their negated vectors) W(x,y,z) <0 if and only if W(-x,-y,-z) > 0. Lecture 1 35
  • 36. Proof (cont).  Consider quotient V*W/|W|. (note: this is multidimensional cosine between V* and W.) Recall V* is unit vector . Quotient <= 1. Lecture 1 36
  • 37. Proof(cont) Now each time FIX is visited W changes via ADD. V* W(n+1) = V*(W(n) + X) = V* W(n) + V*X >= V* W(n) + δ Hence V* W(n) >= n δ (∗) Lecture 1 37
  • 38. Proof (cont)  Now consider denominator:  |W(n+1)| = W(n+1)W(n+1) = ( W(n) + X)(W(n) + X) = |W(n)| **2 + 2W(n)X + 1 (recall |X| = 1 and W(n)X < 0 since X is positive example and we are in FIX) < |W(n)|**2 + 1 So after n times |W(n+1)|**2 < n (**) Lecture 1 38
  • 39. Proof (cont)  Putting (*) and (**) together: Quotient = V*W/|W| > nδ/ sqrt(n) Since Quotient <=1 this means n < (1/δ)∗∗2. This means we enter FIX a bounded number of times. Q.E.D. Lecture 1 39
  • 40. Geometric Proof  See hand slides. Lecture 1 40
  • 41. Perceptron Diagram 1 Solution OK No examples No examples No examples Lecture 1 41
  • 42. Solution OK No examples No examples No examples Lecture 1 42
  • 43. Lecture 1 43
  • 44. Additional Facts  Note: If X’s presented in systematic way, then solution W always found.  Note: Not necessarily same as V*  Note: If F not finite, may not obtain solution in finite time  Can modify algorithm in minor ways and stays valid (e.g. not unit but bounded examples); changes in W(n). Lecture 1 44
  • 45. Perceptron Convergence Theorem  If the concept is representable in a perceptron then the perceptron learning rule will converge in a finite amount of time.  (MAKE PRECISE and Prove) Lecture 1 45
  • 46. Important Points  Theorem only guarantees result IF representable!  Usually we are not interested in just representing something but in its generalizability – how will it work on examples we havent seen! Lecture 1 46
  • 47. Percentage of Boolean Functions Representible by a Perceptron  Input Perceptron Functions 1 4 4 2 16 14 3 104 256 4 1,882 65,536 5 94,572 10**9 6 15,028,134 10**19 7 8,378,070,864 10**38 8 17,561,539,552,946 10**77 Lecture 1 47
  • 48. Generalizability  Typically train a network on a sample set of examples  Use it on general class  Training can be slow; but execution is fast. Lecture 1 48
  • 49. Perceptron •weights ∑w x i i > threshold ⇔ A in receptive field kdkdkfjlll ∑w x i i > θ ⇔ The letter A is in the recept •Pattern Identification •(Note: Neuron is trained) Lecture 1 49
  • 50. What wont work?  Example: Connectedness with bounded diameter perceptron.  Compare with Convex with (use sensors of order three). Lecture 1 50
  • 51. Biases and Thresholds  We can replace the threshold with a bias. n ∑ wx >θ i i  A bias acts i= 1 n exactly as a weight on a ∑ w x -θ > 0 i= 1 i i n connection from a unit ∑ wx > 0 i= 0 i i w0 = - θ x0 = 1 whose activation is always 1. Lecture 1 51
  • 52. Perceptron  Loop :  Take an example and apply to network.  If correct answer – return to Loop.  If incorrect – go to Fix.  Fix :  Adjust network weights by input example.  Go to Loop. Lecture 1 52
  • 53. Perceptron Algorithm v x∈ F Let be arbitrary Choose: chooseF Test: v ⋅Ifx > 0x∈ x ∈ and F+ go to v⋅ x ≤ Choose 0 x∈ F + If and go to v⋅ x ≤ 0 Fix plus x∈ F − vIf > 0 ⋅x x ∈ and F− go to v := v + x Choose If and go to v := v − x Fix minus Fix plus: go to Choose Fix minus: go to Choose Lecture 1 53
  • 54. Perceptron Algorithm  Conditions to the algorithm existence : Condition no.1: +    if x ∈ F v * ⋅ x > 0 ∃ v∗ ,   if x ∈ F v * ⋅ x ≤ 0 −  Condition no.2: We choose F to be a group of unit vectors. Lecture 1 54
  • 55. Geometric viewpoint wn + 1 x v * wn Lecture 1 55
  • 56. Perceptron Algorithm  Based on these conditions the number of times we enter the Loop is finite. Examples  Proof: world + − F = F F { x A x > 0} { y A y ≤ 0} * * Positiv Negati e ve exampl examp es les Lecture 1 56
  • 57. Perceptron Algorithm- Proof  We replace the threshold with a bias.  We assume F is a group of unit vectors. ∀φ ∈ F φ =1 Lecture 1 57
  • 58. Perceptron Algorithm- Proof We reduce ∑ ay≤0 ∑ a (− y ) ≥ 0  what we have * i i ⇒ * i i to prove by yi ⇒ -yi eliminating all + the negative F= F examples and placing their ∀φ ∈ F A*φ > 0 negations in the positive ∀φ ∈ F A*φ > δ examples. ⇓ ∃δ > 0 Aφ > δ * Lecture 1 58
  • 59. Perceptron Algorithm- Proof A* ⋅ A A* ⋅ A G ( A) = * = ≤1 A ⋅A A The numerator : A ⋅ Al + 1 = A ⋅ ( Al + φ ) = A ⋅ Al + A ⋅ φ ≥ A ⋅ Al + δ * * * * * >δ After n changes A* ⋅ An ≥ n ⋅ δ Lecture 1 59
  • 60. Perceptron Algorithm- Proof The denominator : At + 1 = ( At + 1 )( At + 1 ) = ( At + φ )( At + φ ) = 2 2 2 2 = At + 2 ⋅ At ⋅ φ + φ < At + 1 <0 =1 After n 2 changes An < n Lecture 1 60
  • 61. Perceptron Algorithm- Proof From the denominato r: 2 An < n From the numerator : A* ⋅ An > n ⋅ δ A ⋅ An n ⋅ δ * G ( A) = > = n ⋅δ ≤ 1 An n ⇓ n is 1 final n< δ 2 Lecture 1 61
  • 62. Example - AND x1 x2 bias AND w0 = 0,0,0 0 w0 ⋅ x 0 = w0 ⋅ 0,0,1 = 0 w0 ⋅ x1 = w0 ⋅ 0,1,1 = 0 0 0 1 0 w0 ⋅ x 2 = w0 ⋅ 1,0,1 = 0  wrong 0 1 1 0 w0 ⋅ x3 = w0 ⋅ 1,1,1 = 0 FIX = w1 = w0 + x3 = 0,0,0 + 1,1,1 = 1,1,1 1 0 1 1 w1 ⋅ x 0 = w1 ⋅ 0,0,1 = 1 FIX = w2 = w1 − x 0 = 1,1,1 − 0,0,1wrong  = 1,1,0 1 1 1 w2 ⋅ x1 = w2 ⋅ 0,1,1 = 1 FIX = w3 = w2 − x1 = 1,1,0 − 0,1,1 = 1,0,−1  wrong …etc Lecture 1 62
  • 63. AND – Bi Polar - solution - 1 0 - - 1 1 1 0 1 1 - - 1 1 0 1 1 0 1 1 - - 1 1 0 1 1 0 1 1 1 1 1 0  wrong+ 1 1 1 1 - - 1 1 1 1 0 1 1 - 1 1 1 1  wrong - success - 0 2 0 1 1 1 1  wrong - 1 1 - 1 1 1 1 1 continue Lecture 1 63
  • 64. Problem LHS1 MHS1 RHS1 + MHS2 MHS 2RHS < LHSLHS 3 + 3MHS 33 + RHS 3 > 0 LHS 2 LHS + RHS 2 0 3 MHS RHS 2 2 LHS1 + MHS1 + RHS1 > 0 MHS1 = MHS 2 = MHS 3 LHS1 + RHS1 > 0 LHS 2 + RHS 2 < 0 LHS 3 + RHS3 > 0 LHS1 = LHS 2 RHS 2 = RHS 3 should be small RHS 2 should be small LHS 2 enough so that enough so that LHS 2 + RHS 2 < 0 LHS 2 + RHS 2 < 0 contradictio !!!n Lecture 1 64
  • 65. Linear Separation  Every perceptron determines a classification of vector inputs which is determined by a hyperline  Two dimensional examples (add algebra) OR AND XOR not possible Lecture 1 65
  • 66. Linear Separation in Higher Dimensions In higher dimensions, still linear separation, but hard to tell Example: Connected; Convex - which can be handled by Perceptron with local sensors; which can not be. Note: Define local sensors. Lecture 1 66
  • 67. What wont work?  Try XOR. Lecture 1 67
  • 68. Limitations of Perceptron  Representability  Only concepts that are linearly separable.  Compare: Convex versus connected  Examples: XOR vs OR Lecture 1 68