Original SOINN

SOINN Inc.
SOINN Inc.像情報工学研究所 長谷川研究室 at SOINN Inc.
An incremental network for on-line
 unsupervised classification and
        topology learning

      Shen Furao       Osamu Hasegawa


Neural Networks, Vol.19, No.1, pp.90-106, (2006)
Background: Objective of unsupervised learning (1)
    Clustering: Construct decision boundaries
      based on unlabeled data.
      – Single-link, complete-link, CURE
         • Computation overload
         • Much memory space
         • Unsuitable for large data sets or online data
      – K-means:
         •   Dependence on initial starting conditions
         •   Tendency to result in local minima
         •   Determine the number of clusters k in advance
         •   data sets consisting only of isotropic clusters
Background: Objective of unsupervised learning (2)
Topology learning: Given some high-dimensional data
  distribution, find a topological structure that closely
  reflects the topology of the data distribution
   – SOM: self-organizing map
       • predetermined structure and size
       • posterior choice of class labels for the prototypes
   – CHL+NG: competitive Hebbian learning + neural gas
       • a priori decision about the network size
       • ranking of all nodes in each adaptation step
       • use of adaptation parameter
   – GNG: growing neural gas
       • permanent increase in the number of nodes
       • permanent drift of centers to capture input probability
         density
Background: Online or life-long learning
Fundamental issue (Stability-Plasticity Dilemma): How can
  a learning system adapt to new information without
  corrupting or forgetting previously learned information
   – GNG-U: deletes nodes which are located in regions of
     a low input probability density
      • learned old prototype patterns will be destroyed
   – Hybrid network: Fuzzy ARTMAP + PNN
   – Life-long learning with improved GNG: learn number
     of nodes needed for current task
      • only for supervised life-long learning
Objectives of proposed algorithm
• To process the on-line non-stationary data.
• To do the unsupervised learning without any priori
  condition such as:
   • suitable number of nodes
   • a good initial codebook
   • how many classes there are
• Report a suitable number of classes
• Represent the topological structure of the input
  probability density.
• Separate the classes with some low-density overlaps
• Detect the main structure of clusters polluted by noises.
Proposed algorithm

          First Layer            Second Layer
Input      Growing       First     Growing        Second
pattern    Network      Output     Network        Output




          Insert        Delete
                                       Classify
          Node          Node
Algorithms
• Insert new nodes
  – Criterion: nodes with high errors serve as a criterion
    to insert a new node
  – error-radius is used to judge if the insert is successful
• Delete nodes
  – Criterion: remove nodes in low probability density
    regions
  – Realize: delete nodes with no or only one direct
    topology neighbor
• Classify
  – Criterion: all nodes linked with edges will be one
    cluster
Experiment
• Stationary environment: patterns are randomly chosen
  from all area A, B, C, D and E
• NON-Stationary environment:
                                       Environment
                                 I II III IV V VI VII
                            A 1 0 1 0 0 0 0
                             B 0 1 0 1 0 0 0
                             C 0 0 1 0 0 1 0
                            D 0 0 0 1 1 0 0
                            E1 0 0 0 0 1 0 0
                            E2 0 0 0 0 0 1 0
   Original Data Set        E3 0 0 0 0 0 0 1
Experiment: Stationary environment




Original Data Set   Traditional method: GNG
Experiment: Stationary environment




Proposed method: first layer   Proposed method: final results
Experiment: Non-stationary environment




  GNG result              GNG-U result
Experiment: Non-stationary environment




         Proposed method: first layer
Experiment: Non-stationary environment




         Proposed method: first layer
Experiment: Non-stationary environment




Proposed method: first layer Proposed method: Final output
Experiment: Non-stationary environment




Number of growing nodes during online learning
     (Environment 1 ~ Environment 7)
Experiment: Real World Data
(ATT_FACE)
Facial Image




               (a) 10 classes




               (b) 10 samples of class 1
Experiment:Vector




           Vector of (a)




           Vector of (b)
Experiment: Face Recognition results
                             10 clusters

                             Stationary
                             Correct
                             Recognition
                             Ratio: 90%

                             Non-Stationary
                             Correct
                             Recognition
                             Ratio: 86%
Experiment: Vector Quantization




                             Stationary Environment: Decoding
 Original Lena (512*512*8)
                             image, 130 nodes, 0.45bpp,
                             PSNR = 30.79dB
Experiment: Compare with GNG
   Stationary Environment

                      Number
                                  bpp    PSNR
                      of Nodes
      First-layer           130   0.45   30.79

         GNG                130   0.45   29.98

     Second-layer           52    0.34   29.29

         GNG                52    0.34   28.61
Experiment: Non-stationary Environment




First-layer: 499 nodes, 0.56bpp,   Second-layer: 64 nodes, 0.375bpp,
PSNR = 32.91dB                     PSNR = 29.66dB
Conclusion
• An autonomous learning system for
  unsupervised classification and topology
  representation task
• Grow incrementally and learn the number of
  nodes needed to solve current task
• Accommodate input patterns of on-line non-
  stationary data distribution
• Eliminate noise in the input data
1 of 22

More Related Content

What's hot(20)

Deep Learning and TensorFlowDeep Learning and TensorFlow
Deep Learning and TensorFlow
Oswald Campesato118 views
Skip RNN: Learning to Skip State Updates in Recurrent Neural NetworksSkip RNN: Learning to Skip State Updates in Recurrent Neural Networks
Skip RNN: Learning to Skip State Updates in Recurrent Neural Networks
Universitat Politècnica de Catalunya1.5K views
Yolo releases gianmariaYolo releases gianmaria
Yolo releases gianmaria
Deep Learning Italia389 views
The Perceptron - Xavier Giro-i-Nieto - UPC Barcelona 2018The Perceptron - Xavier Giro-i-Nieto - UPC Barcelona 2018
The Perceptron - Xavier Giro-i-Nieto - UPC Barcelona 2018
Universitat Politècnica de Catalunya679 views
Backpropagation - Elisa Sayrol - UPC Barcelona 2018Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Universitat Politècnica de Catalunya469 views

Viewers also liked(12)

NDVINDVI
NDVI
Elizabeth Wesley1.4K views
ClusteringClustering
Clustering
M Rizwan Aqeel21.8K views
Supervised ClassificationSupervised Classification
Supervised Classification
Chad Yowler1.9K views
Presentacion elafis Brasil 2009Presentacion elafis Brasil 2009
Presentacion elafis Brasil 2009
Fredy Neira520 views
Pert & CpmPert & Cpm
Pert & Cpm
Vikram Thadeshvar45.5K views
Program Evaluation and Review Technique (PERT)Program Evaluation and Review Technique (PERT)
Program Evaluation and Review Technique (PERT)
Abhishek Pachisia33.8K views
Pert & cpm   project managementPert & cpm   project management
Pert & cpm project management
Rahul Dubey165.3K views

Similar to Original SOINN(20)

More from SOINN Inc.

PBAIPBAI
PBAISOINN Inc.
28.4K views28 slides
Dr.Kawewong Ph.D ThesisDr.Kawewong Ph.D Thesis
Dr.Kawewong Ph.D ThesisSOINN Inc.
8.8K views48 slides
II
ISOINN Inc.
6.9K views23 slides
PIRF-NAV2PIRF-NAV2
PIRF-NAV2SOINN Inc.
8.4K views32 slides
SSA-SOINNSSA-SOINN
SSA-SOINNSOINN Inc.
40.4K views25 slides
E-SOINNE-SOINN
E-SOINNSOINN Inc.
41.3K views27 slides

More from SOINN Inc.(14)

PBAIPBAI
PBAI
SOINN Inc.28.4K views
Dr.Kawewong Ph.D ThesisDr.Kawewong Ph.D Thesis
Dr.Kawewong Ph.D Thesis
SOINN Inc.8.8K views
II
I
SOINN Inc.6.9K views
PIRF-NAV2PIRF-NAV2
PIRF-NAV2
SOINN Inc.8.4K views
SSA-SOINNSSA-SOINN
SSA-SOINN
SOINN Inc.40.4K views
E-SOINNE-SOINN
E-SOINN
SOINN Inc.41.3K views
SOINN PBRSOINN PBR
SOINN PBR
SOINN Inc.5K views
SOIAM (SOINN-AM)SOIAM (SOINN-AM)
SOIAM (SOINN-AM)
SOINN Inc.1.6K views
SOINN-AMSOINN-AM
SOINN-AM
SOINN Inc.10.4K views
学生さんへのメッセージ学生さんへのメッセージ
学生さんへのメッセージ
SOINN Inc.661 views

Recently uploaded(20)

Green Leaf Consulting: Capabilities DeckGreen Leaf Consulting: Capabilities Deck
Green Leaf Consulting: Capabilities Deck
GreenLeafConsulting170 views
METHOD AND SYSTEM FOR PREDICTING OPTIMAL LOAD FOR WHICH THE YIELD IS MAXIMUM ...METHOD AND SYSTEM FOR PREDICTING OPTIMAL LOAD FOR WHICH THE YIELD IS MAXIMUM ...
METHOD AND SYSTEM FOR PREDICTING OPTIMAL LOAD FOR WHICH THE YIELD IS MAXIMUM ...
Prity Khastgir IPR Strategic India Patent Attorney Amplify Innovation23 views
Java Platform Approach 1.0 - Picnic MeetupJava Platform Approach 1.0 - Picnic Meetup
Java Platform Approach 1.0 - Picnic Meetup
Rick Ossendrijver23 views
CXL at OCPCXL at OCP
CXL at OCP
CXL Forum183 views
Liqid: Composable CXL PreviewLiqid: Composable CXL Preview
Liqid: Composable CXL Preview
CXL Forum118 views

Original SOINN

  • 1. An incremental network for on-line unsupervised classification and topology learning Shen Furao Osamu Hasegawa Neural Networks, Vol.19, No.1, pp.90-106, (2006)
  • 2. Background: Objective of unsupervised learning (1) Clustering: Construct decision boundaries based on unlabeled data. – Single-link, complete-link, CURE • Computation overload • Much memory space • Unsuitable for large data sets or online data – K-means: • Dependence on initial starting conditions • Tendency to result in local minima • Determine the number of clusters k in advance • data sets consisting only of isotropic clusters
  • 3. Background: Objective of unsupervised learning (2) Topology learning: Given some high-dimensional data distribution, find a topological structure that closely reflects the topology of the data distribution – SOM: self-organizing map • predetermined structure and size • posterior choice of class labels for the prototypes – CHL+NG: competitive Hebbian learning + neural gas • a priori decision about the network size • ranking of all nodes in each adaptation step • use of adaptation parameter – GNG: growing neural gas • permanent increase in the number of nodes • permanent drift of centers to capture input probability density
  • 4. Background: Online or life-long learning Fundamental issue (Stability-Plasticity Dilemma): How can a learning system adapt to new information without corrupting or forgetting previously learned information – GNG-U: deletes nodes which are located in regions of a low input probability density • learned old prototype patterns will be destroyed – Hybrid network: Fuzzy ARTMAP + PNN – Life-long learning with improved GNG: learn number of nodes needed for current task • only for supervised life-long learning
  • 5. Objectives of proposed algorithm • To process the on-line non-stationary data. • To do the unsupervised learning without any priori condition such as: • suitable number of nodes • a good initial codebook • how many classes there are • Report a suitable number of classes • Represent the topological structure of the input probability density. • Separate the classes with some low-density overlaps • Detect the main structure of clusters polluted by noises.
  • 6. Proposed algorithm First Layer Second Layer Input Growing First Growing Second pattern Network Output Network Output Insert Delete Classify Node Node
  • 7. Algorithms • Insert new nodes – Criterion: nodes with high errors serve as a criterion to insert a new node – error-radius is used to judge if the insert is successful • Delete nodes – Criterion: remove nodes in low probability density regions – Realize: delete nodes with no or only one direct topology neighbor • Classify – Criterion: all nodes linked with edges will be one cluster
  • 8. Experiment • Stationary environment: patterns are randomly chosen from all area A, B, C, D and E • NON-Stationary environment: Environment I II III IV V VI VII A 1 0 1 0 0 0 0 B 0 1 0 1 0 0 0 C 0 0 1 0 0 1 0 D 0 0 0 1 1 0 0 E1 0 0 0 0 1 0 0 E2 0 0 0 0 0 1 0 Original Data Set E3 0 0 0 0 0 0 1
  • 9. Experiment: Stationary environment Original Data Set Traditional method: GNG
  • 10. Experiment: Stationary environment Proposed method: first layer Proposed method: final results
  • 11. Experiment: Non-stationary environment GNG result GNG-U result
  • 12. Experiment: Non-stationary environment Proposed method: first layer
  • 13. Experiment: Non-stationary environment Proposed method: first layer
  • 14. Experiment: Non-stationary environment Proposed method: first layer Proposed method: Final output
  • 15. Experiment: Non-stationary environment Number of growing nodes during online learning (Environment 1 ~ Environment 7)
  • 16. Experiment: Real World Data (ATT_FACE) Facial Image (a) 10 classes (b) 10 samples of class 1
  • 17. Experiment:Vector Vector of (a) Vector of (b)
  • 18. Experiment: Face Recognition results 10 clusters Stationary Correct Recognition Ratio: 90% Non-Stationary Correct Recognition Ratio: 86%
  • 19. Experiment: Vector Quantization Stationary Environment: Decoding Original Lena (512*512*8) image, 130 nodes, 0.45bpp, PSNR = 30.79dB
  • 20. Experiment: Compare with GNG Stationary Environment Number bpp PSNR of Nodes First-layer 130 0.45 30.79 GNG 130 0.45 29.98 Second-layer 52 0.34 29.29 GNG 52 0.34 28.61
  • 21. Experiment: Non-stationary Environment First-layer: 499 nodes, 0.56bpp, Second-layer: 64 nodes, 0.375bpp, PSNR = 32.91dB PSNR = 29.66dB
  • 22. Conclusion • An autonomous learning system for unsupervised classification and topology representation task • Grow incrementally and learn the number of nodes needed to solve current task • Accommodate input patterns of on-line non- stationary data distribution • Eliminate noise in the input data