SlideShare a Scribd company logo
1 of 13
Representation of object position in various frames
of reference using robotic simulator

Gain Modulation

CNC seminar                               5.12.2012



Student:    Marcel Švec
Supervisor: doc. Ing. Igor Farkaš, PhD.
2


Annotation
• Humans use several egocentric frames of reference for effective sensomotoric

 coordination (eye-hand): eye-centered, head-centered, body-centered location…

• How does brain carry out computations such as coordinate transformation? How

 are sensory inputs translated to motor outputs?

• Our aim is to implement and evaluate neural network model capable of

 coordinate translation

• Goal of the design is to create inner representation of the surrounding space in

 simulated agent

• We use robotic simulator iCub for experiments


• Study of area 7a (visual and eye-position neurons), most likely this area

 preforms spatial transformations
3


Experiment One – introduction
• Inspired by the article:
  • A back-propagation programmed network that simulates response
    properties of a subset of posterior parietal neurons. David Zipser,
    Richard A. Andersen (Nature, 1988) [http://dx.doi.org/10.1038/331679a0]



• Excerpt:

  • Experiments in macaque monkeys (proposed that
    area 7a contains visual and eye-position neurons)

  • Experimental results from area 7a:
    eye-position neurons (15%),
    visual neurons (21%), combination (57%)

  • Interaction between eye-position and visual
    responses ( gain fields)                                          Image source:
                                                                       http://en.wikipedia.org/wiki/Brodmann_area_7
4


Experiment One – spatial gain fields
• Determining the effect of eye position on receptive fields
  • visual stimulus always present at the same retinal location
  • monkey has had fixed and fixates on a point f




                                                       (Zipser, Andersen 1988)
5


Experiment One – neural network model
Proposed neural network model:

• Three layer, back-propagation

• Input: visual stimulus and eye
  position

• Output: head-centered coordinates




                                      Model and experimental retinal
                                      receptive fields were remarkably
                                      similar               (Zipser, Andersen 1988)
6


Experiment One – dataset
• Input:
      • eye position – vertical and horizontal orientation (angle)
      • visual stimulus – images from the left and right eye (processed)



• Output:
  • Body-referenced target
       position expressed by
       horizontal and vertical
       slope


• iCub generator:
  •    controlling iCub (eyes limits)
  •    where to put an object (camera
       properties)
  •    eyes rotation (keeping object in
       FOV)
  •    what size (scaling)
  •    processing (color filter)
7


Experiment One – training, testing, results
• FANN –        Fast Artificial Neural Network Library
  •   C, multilayer neural networks, back-propagation training (incremental, RPROP, Quickprop, batch), cross-platform,
      bindings to >15 languages…



• Network model:
  • Input layer = eye_tilt + eye_version + left_eye_image + right_eye_image = 11 + 21 + 64*48 + 64*48 = 6176 neurons
  • Hidden layer = 250 neurons
  • Output layer = x-slope + y-slope = 10 + 10 = 20 neurons (every 20 degrees)



• So far best results:
  • Sigmoids, steepness hidden = 0.05 (1/20), output = 0.0715 (1/14)
  • Algorithm – RPROP (batch, doesn’t use learning rate)
  • 93 epochs on 1000 patterns, MSE < 0.0001, training took less than 15 minutes,
      comparing to incremental training – 200 epochs, MSE < 0.0017, 53 minutes
  • Average error: x < 4, y < 3.3 degrees, standard deviations: x = 3.6, y = 3.2
8


    Experiment One – results




                                                             sorted errors on testing data


                                                      •   Possible cause of inaccuracies is
                                                          the character of patterns – objects
                                                          of different size, shape and
                                                          orientation may appear at the same
                                                          location, which means that inputs
                                                          that differ in visual stimulus
                                                          (size, shading) may require the
                                                          same output.

                                                      •   Solutions (open): another hidden
                                                          layer, different network model?
Comparing real and desired values of output neurons
9


Gain fields
Mapping the receptive
field of parietal neuron

Stimulus is present at
the same retinal location   (head turned left)



Neural response of the
neuron when the head is
turned right and left:

• location and shape
  remains

• amplitude (gain)
  changes



                                 Salinas et al. (2001)
10


Gain fields – computing
• Gain fields of parietal neurons depend on eye position, head position…
  note: although single neurons are modulated, several neurons have to be combined to get the whole information
  (population code)



• r = f(x   – a) g(y)                                            // single parietal neuron
   • x      – retinal location of the stimulus
   • y      – gaze angle
   • f      – function for the response to the visual stimulus
   • a      – location of the peak of function f
   • r      – amplitude of the response
   • g      – gain field


• R = F(c1x + c2y)                                               // downstream response
   • F      – peaked function that represent receptive field

• Set of downstream neurons may represent quantity (x + y), while another (x-y), both sets being driven
  by the same population of gain-modulated neurons.
• Very general mechanism.
                                                                                                   Salinas et al. (2001)
11


Gain modulation and coordinate transformations
• Gain modulated neurons are suited to add, subtract and performs operations essential for coordinate
  transformations.


• Figure:


   • 3 eyes positions


   • left columns –
     responses of 4 idealized
     gain-modulated neurons

   • right column –
     responses of
     a downstream neuron
     (weighted sum)




                                                                                      Salinas et al. (2001)
12


Gain modulation
• it is an extremely widespread mechanism

• non-linear combination of information from several sources

• affected is sensitivity (amplitude, gain), not selectivity (sensitivity or
 receptive field properties)
• indications that it serves as a basis for computations (coordinate
 transformations, invariant responses)


• related also to
  • Object recognition (invariance)

  • Focusing attention

  • Motion processing
13




                                                        http://masterthesis.itbrutus.com




                           Thanks for your attention



Cited sources:

•   Salinas, E. and T. J. Sejnowski (2001). Gain modulation in the central nervous system: Where
    behavior, neurophysiology, and computation meet. The Neuroscientist 7, pp. 430440.

•   Zipser, D. and R. A. Andersen (1988). A backpropagation programmed network that simulates
    response properties of a subset of posterior parietal neurons. Nature 331, pp. 679684.

More Related Content

What's hot (7)

Contribution_of_the_polarimetric_information.pdf
Contribution_of_the_polarimetric_information.pdfContribution_of_the_polarimetric_information.pdf
Contribution_of_the_polarimetric_information.pdf
 
Jennie Si: "Computing with Neural Spikes"
Jennie Si: "Computing with Neural Spikes" Jennie Si: "Computing with Neural Spikes"
Jennie Si: "Computing with Neural Spikes"
 
Inter-observers congruency and memorability
Inter-observers congruency and memorabilityInter-observers congruency and memorability
Inter-observers congruency and memorability
 
Characterizing the aberrations in fluorescence macroscope
Characterizing the aberrations in fluorescence macroscopeCharacterizing the aberrations in fluorescence macroscope
Characterizing the aberrations in fluorescence macroscope
 
2009 0359ogidigbe Arvo Final
2009 0359ogidigbe Arvo Final2009 0359ogidigbe Arvo Final
2009 0359ogidigbe Arvo Final
 
Critical Features for Recognition of Biological Motion
Critical Features for Recognition of Biological MotionCritical Features for Recognition of Biological Motion
Critical Features for Recognition of Biological Motion
 
Guided tour of visual attention
Guided tour of visual attentionGuided tour of visual attention
Guided tour of visual attention
 

Similar to Short presentation about my thesis

2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…
2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…
2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…
Dongseo University
 
Fcv bio cv_cottrell
Fcv bio cv_cottrellFcv bio cv_cottrell
Fcv bio cv_cottrell
zukun
 
Fcv bio cv_cottrell
Fcv bio cv_cottrellFcv bio cv_cottrell
Fcv bio cv_cottrell
zukun
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
Hưng Đặng
 
Jennie sinsfadp06
Jennie sinsfadp06Jennie sinsfadp06
Jennie sinsfadp06
sundarnu
 

Similar to Short presentation about my thesis (20)

Neural fields, a cognitive approach
Neural fields, a cognitive approachNeural fields, a cognitive approach
Neural fields, a cognitive approach
 
Neural networks
Neural networksNeural networks
Neural networks
 
L005.neural networks
L005.neural networksL005.neural networks
L005.neural networks
 
2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…
2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…
2013-1 Machine Learning Lecture 04 - Torsten Reil - Artificial Neural Netw…
 
Neural net and back propagation
Neural net and back propagationNeural net and back propagation
Neural net and back propagation
 
Object Tracking using Artificial Neural Network
Object Tracking using Artificial Neural NetworkObject Tracking using Artificial Neural Network
Object Tracking using Artificial Neural Network
 
Fcv bio cv_cottrell
Fcv bio cv_cottrellFcv bio cv_cottrell
Fcv bio cv_cottrell
 
Fcv bio cv_cottrell
Fcv bio cv_cottrellFcv bio cv_cottrell
Fcv bio cv_cottrell
 
Recent advances of AI for medical imaging : Engineering perspectives
Recent advances of AI for medical imaging : Engineering perspectivesRecent advances of AI for medical imaging : Engineering perspectives
Recent advances of AI for medical imaging : Engineering perspectives
 
Facial emotion detection on babies' emotional face using Deep Learning.
Facial emotion detection on babies' emotional face using Deep Learning.Facial emotion detection on babies' emotional face using Deep Learning.
Facial emotion detection on babies' emotional face using Deep Learning.
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
Lecture artificial neural networks and pattern recognition
Lecture   artificial neural networks and pattern recognitionLecture   artificial neural networks and pattern recognition
Lecture artificial neural networks and pattern recognition
 
Introduction to Artificial Neural Networks
Introduction to Artificial Neural Networks Introduction to Artificial Neural Networks
Introduction to Artificial Neural Networks
 
Neural Networks-1
Neural Networks-1Neural Networks-1
Neural Networks-1
 
Facebook Deep face
Facebook Deep faceFacebook Deep face
Facebook Deep face
 
Visual brain
Visual brainVisual brain
Visual brain
 
Neural Networks and Deep Learning Basics
Neural Networks and Deep Learning BasicsNeural Networks and Deep Learning Basics
Neural Networks and Deep Learning Basics
 
Neural networks
Neural networksNeural networks
Neural networks
 
Jennie sinsfadp06
Jennie sinsfadp06Jennie sinsfadp06
Jennie sinsfadp06
 
Nv presentation to doctors desk
Nv presentation to doctors deskNv presentation to doctors desk
Nv presentation to doctors desk
 

Recently uploaded

Salient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsSalient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functions
KarakKing
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
heathfieldcps1
 

Recently uploaded (20)

Single or Multiple melodic lines structure
Single or Multiple melodic lines structureSingle or Multiple melodic lines structure
Single or Multiple melodic lines structure
 
How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17How to Create and Manage Wizard in Odoo 17
How to Create and Manage Wizard in Odoo 17
 
Understanding Accommodations and Modifications
Understanding  Accommodations and ModificationsUnderstanding  Accommodations and Modifications
Understanding Accommodations and Modifications
 
Salient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functionsSalient Features of India constitution especially power and functions
Salient Features of India constitution especially power and functions
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Graduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - EnglishGraduate Outcomes Presentation Slides - English
Graduate Outcomes Presentation Slides - English
 
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
Sensory_Experience_and_Emotional_Resonance_in_Gabriel_Okaras_The_Piano_and_Th...
 
General Principles of Intellectual Property: Concepts of Intellectual Proper...
General Principles of Intellectual Property: Concepts of Intellectual  Proper...General Principles of Intellectual Property: Concepts of Intellectual  Proper...
General Principles of Intellectual Property: Concepts of Intellectual Proper...
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
Food safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdfFood safety_Challenges food safety laboratories_.pdf
Food safety_Challenges food safety laboratories_.pdf
 
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
 
How to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POSHow to Manage Global Discount in Odoo 17 POS
How to Manage Global Discount in Odoo 17 POS
 
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
80 ĐỀ THI THỬ TUYỂN SINH TIẾNG ANH VÀO 10 SỞ GD – ĐT THÀNH PHỐ HỒ CHÍ MINH NĂ...
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 

Short presentation about my thesis

  • 1. Representation of object position in various frames of reference using robotic simulator Gain Modulation CNC seminar 5.12.2012 Student: Marcel Švec Supervisor: doc. Ing. Igor Farkaš, PhD.
  • 2. 2 Annotation • Humans use several egocentric frames of reference for effective sensomotoric coordination (eye-hand): eye-centered, head-centered, body-centered location… • How does brain carry out computations such as coordinate transformation? How are sensory inputs translated to motor outputs? • Our aim is to implement and evaluate neural network model capable of coordinate translation • Goal of the design is to create inner representation of the surrounding space in simulated agent • We use robotic simulator iCub for experiments • Study of area 7a (visual and eye-position neurons), most likely this area preforms spatial transformations
  • 3. 3 Experiment One – introduction • Inspired by the article: • A back-propagation programmed network that simulates response properties of a subset of posterior parietal neurons. David Zipser, Richard A. Andersen (Nature, 1988) [http://dx.doi.org/10.1038/331679a0] • Excerpt: • Experiments in macaque monkeys (proposed that area 7a contains visual and eye-position neurons) • Experimental results from area 7a: eye-position neurons (15%), visual neurons (21%), combination (57%) • Interaction between eye-position and visual responses ( gain fields) Image source: http://en.wikipedia.org/wiki/Brodmann_area_7
  • 4. 4 Experiment One – spatial gain fields • Determining the effect of eye position on receptive fields • visual stimulus always present at the same retinal location • monkey has had fixed and fixates on a point f (Zipser, Andersen 1988)
  • 5. 5 Experiment One – neural network model Proposed neural network model: • Three layer, back-propagation • Input: visual stimulus and eye position • Output: head-centered coordinates Model and experimental retinal receptive fields were remarkably similar (Zipser, Andersen 1988)
  • 6. 6 Experiment One – dataset • Input: • eye position – vertical and horizontal orientation (angle) • visual stimulus – images from the left and right eye (processed) • Output: • Body-referenced target position expressed by horizontal and vertical slope • iCub generator: • controlling iCub (eyes limits) • where to put an object (camera properties) • eyes rotation (keeping object in FOV) • what size (scaling) • processing (color filter)
  • 7. 7 Experiment One – training, testing, results • FANN – Fast Artificial Neural Network Library • C, multilayer neural networks, back-propagation training (incremental, RPROP, Quickprop, batch), cross-platform, bindings to >15 languages… • Network model: • Input layer = eye_tilt + eye_version + left_eye_image + right_eye_image = 11 + 21 + 64*48 + 64*48 = 6176 neurons • Hidden layer = 250 neurons • Output layer = x-slope + y-slope = 10 + 10 = 20 neurons (every 20 degrees) • So far best results: • Sigmoids, steepness hidden = 0.05 (1/20), output = 0.0715 (1/14) • Algorithm – RPROP (batch, doesn’t use learning rate) • 93 epochs on 1000 patterns, MSE < 0.0001, training took less than 15 minutes, comparing to incremental training – 200 epochs, MSE < 0.0017, 53 minutes • Average error: x < 4, y < 3.3 degrees, standard deviations: x = 3.6, y = 3.2
  • 8. 8 Experiment One – results sorted errors on testing data • Possible cause of inaccuracies is the character of patterns – objects of different size, shape and orientation may appear at the same location, which means that inputs that differ in visual stimulus (size, shading) may require the same output. • Solutions (open): another hidden layer, different network model? Comparing real and desired values of output neurons
  • 9. 9 Gain fields Mapping the receptive field of parietal neuron Stimulus is present at the same retinal location (head turned left) Neural response of the neuron when the head is turned right and left: • location and shape remains • amplitude (gain) changes Salinas et al. (2001)
  • 10. 10 Gain fields – computing • Gain fields of parietal neurons depend on eye position, head position… note: although single neurons are modulated, several neurons have to be combined to get the whole information (population code) • r = f(x – a) g(y) // single parietal neuron • x – retinal location of the stimulus • y – gaze angle • f – function for the response to the visual stimulus • a – location of the peak of function f • r – amplitude of the response • g – gain field • R = F(c1x + c2y) // downstream response • F – peaked function that represent receptive field • Set of downstream neurons may represent quantity (x + y), while another (x-y), both sets being driven by the same population of gain-modulated neurons. • Very general mechanism. Salinas et al. (2001)
  • 11. 11 Gain modulation and coordinate transformations • Gain modulated neurons are suited to add, subtract and performs operations essential for coordinate transformations. • Figure: • 3 eyes positions • left columns – responses of 4 idealized gain-modulated neurons • right column – responses of a downstream neuron (weighted sum) Salinas et al. (2001)
  • 12. 12 Gain modulation • it is an extremely widespread mechanism • non-linear combination of information from several sources • affected is sensitivity (amplitude, gain), not selectivity (sensitivity or receptive field properties) • indications that it serves as a basis for computations (coordinate transformations, invariant responses) • related also to • Object recognition (invariance) • Focusing attention • Motion processing
  • 13. 13 http://masterthesis.itbrutus.com Thanks for your attention Cited sources: • Salinas, E. and T. J. Sejnowski (2001). Gain modulation in the central nervous system: Where behavior, neurophysiology, and computation meet. The Neuroscientist 7, pp. 430440. • Zipser, D. and R. A. Andersen (1988). A backpropagation programmed network that simulates response properties of a subset of posterior parietal neurons. Nature 331, pp. 679684.