un supervised learning & self organizing maps. un supervised competitive learning in hebbian...
TRANSCRIPT
Un Supervised Competitive Learning
• In Hebbian networks, all neurons can fire at the same time
• Competitive learning means that only a single neuron from each group fires at each time step
• Output units compete with one another. • These are winner takes all units
(grandmother cells)
UnSupervised Competitive Learning
• In the hebbian like models, all the neurons can fire together
• In Competitive Learning models, only one unit (or one per group) can fire at a time
• Output units compete with one another Winner Takes All units
(“grandmother cells”)
US Competitive, Cntd
• Such networks cluster the data points
• The number of clusters is not predefined but is limited to the number of output units
• Applications include VQ, medical diagnosis, document classification and more
Simple Competitive Learning
x1
x2
xN
W11
W12
W22
WP1
WPN
Y1
Y2
YP
N inputs unitsP output neuronsP x N weights
Pi
N
jjiji XWh
...2,1
1
01oriY
Simple Model, Cntd
• All weights are positive and normalized
• Inputs and outputs are binary
• Only one unit fires in response to an input
)max(arg* i
jijiji
hi
XWXWh
Network Activation
• The unit with the highest field hi fires
• i* is the winner unit
• Geometrically is closest to the current input vector
• The winning unit’s weight vector is updated to be even closer to the current input vector
• Possible variation: adding lateral inhibition
*iW
Learning
Starting with small random weights, at each step:
1. a new input vector is presented to the network
2. all fields are calculated to find a winner
3. is updated to be closer to the input*iW
Learning Rule
• Standard Competitive Learning
)( ** jijji WXW Can be formulated as hebbian:
)( ijjiij WXOW
Competitive Learning, Cntd
• It is important to break the symmetry in the initial random weights
• Final configuration depends on initialization– A winning unit has more chances of winning
the next time a similar input is seen– Some outputs may never fire– This can be compensated by updating the non
winning units with a smaller update
Model: Horizontal & Vertical linesRumelhart & Zipser, 1985
• Problem – identify vertical or horizontal signals
• Inputs are 6 x 6 arrays
• Intermediate layer with 8 WTA units
• Output layer with 2 WTA units
• Cannot work with one layer
Geometrical Interpretation
• So far the ordering of the output units themselves was not necessarily informative
• The location of the winning unit can give us information regarding similarities in the data
• We are looking for an input output mapping that conserves the topologic properties of the inputs feature mapping
• Given any two spaces, it is not guaranteed that such a mapping exits!
Biological Motivation
• In the brain, sensory inputs are represented by topologically ordered computational maps– Tactile inputs – Visual inputs (center-surround, ocular
dominance, orientation selectivity)– Acoustic inputs
Biological Motivation, Cntd
• Computational maps are a basic building block of sensory information processing
• A computational map is an array of neurons representing slightly different tuned processors (filters) that operate in parallel on sensory signals
• These neurons transform input signals into a place coded structure
Self Organizing (Kohonen) Maps
• Competitive networks (WTA neurons)• Output neurons are placed on a lattice, usually 2-
dimensional • Neurons become selectively tuned to various input
patterns (stimuli)• The location of the tuned (winning) neurons become
ordered in such a way that creates a meaningful coordinate system for different input features a topographic map of input patterns is formed
SOMs, Cntd
• Spatial locations of the neurons in the map are indicative of statistical features that are present in the inputs (stimuli)
Self Organization
Kohonen Maps
• Simple case: 2-d input and 2-d output layer
• No lateral connections
• Weight update is done for the winning neuron and its surrounding neighborhood
)*)(,(F ijjij WXiiW
Neighborhood Function
• F is maximal for i* and drops to zero far from i, for example:
• The update “pulls” the winning unit (weight vector) to be closer to the input, and also drags the close neighbors of this unit
)2
exp( F 2
2
*
ii rr
(i,i*)
• The output layer is a sort of an elastic net that wants to come as close as possible to the inputs
• The output maps conserves the topological relationships of the inputs
• Both η and σ can be changed during the learning
Feature Mapping
-6 -4 -2 0 2 4 6 8 10 12 14-4
-3
-2
-1
0
1
2
3
-4 -2 0 2 4 6 8 10 12
-6
-4
-2
0
2
4
6
W(i,1)
W(i,
2)Weight Vectors
-4 -2 0 2 4 6 8 10 12
-6
-4
-2
0
2
4
6
W(i,1)
W(i,
2)Weight Vectors
Topologic Maps in the Brain
• Examples of topologic conserving mapping between input and output spaces– Retintopoical mapping between the retina and
the cortex– Ocular dominance– Somatosensory mapping (the homunculus)
Models
Goodhill (1993) proposed a model for the development of retinotopy and ocular dominance, based on Kohonen Maps– Two retinas project to a single layer of cortical neurons
– Retinal inputs were modeled by random dots patterns
– Added between eyes correlation in the inputs
– The result is an ocular dominance map and a retinotopic map as well
Models, Cntd
Farah (1998) proposed an explanation for the spatial ordering of the homunculus using a simple SOM.– In the womb, the fetus lies with its hands close
to its face, and its feet close to its genitals– This should explain the order of the
somatosensory areas in the homunculus
Other Models
• Semantic self organizing maps to model language acquisition
• Kohonen feature mapping to model layered organization in the LGN
• Combination of unsupervised and supervised learning to model complex computations in the visual cortex
Examples of Applications
• Kohonen (1984). Speech recognition - a map of phonemes in the Finish language
• Optical character recognition - clustering of letters of different fonts
• Angeliol etal (1988) – travelling salesman problem (an optimization problem)
• Kohonen (1990) – learning vector quantization (pattern classification problem)
• Ritter & Kohonen (1989) – semantic maps
Summary• Unsupervised learning is very common • US learning requires redundancy in the stimuli• Self organization is a basic property of the brain’s
computational structure• SOMs are based on
– competition (wta units)– cooperation– synaptic adaptation
• SOMs conserve topological relationships between the stimuli
• Artificial SOMs have many applications in computational neuroscience