Upload
elinor-sharp
View
214
Download
0
Embed Size (px)
Citation preview
2
Unsupervised learning
To-be-learned patterns not wholly provided by modeller
Hebbian unsupervised learning
Competitive learning
4
What’s it good for?
discovering structure in the input
discovering categories in the input Classification networks:
ART (Grossberg & Carpenter)
CALM (Murre & Phaf)
mapping inputs onto a topographic map Kohonen maps (Kohonen) CALM - Maps (Murre & Phaf)
5
Features of Competitive learning Two or more layers (no auto-association)
Competition between output nodes
Two phases: determining a winner learning
Weight normalisation
6
Two or more layers
Input must come from outside the inhibitory clusters
© Rumelhart & Zipser, 1986
7
Competition between output nodes At every presentation of an input pattern,
a winner is determined
Only winner is activated [activation at learning discrete: (0,1) ]
Hard Winner Take All: Find node with maximum input
max. ( wijaj )
Inhibition between nodes
8
Inhibition between nodes
Example: inhibition in CALM
V V
R
V
RR
A
E
Low
HighFlat
Strange
AE
Up Gaussian
Learning intermodular connections
9
Two phases
1. One node wins the competition
2. That node learns, others not
Nodes start off with random weights
No ‘correct’ output connected with inputs: unsupervised learning
10
Weight normalisation
Weights of winner node i changed wij = * aj
Weights add up to constant sum... wij = 1
rule of Rumelhart & Zipser:wij = g * ai / nk - g * wij
…or constant distance: (wij)
2 = 1
11
Geometric interpretation
Both weights & input patterns can be seen as vectors in a hyper space
Euclidian normalisation [ (wij)2 = 1]
all vectors on a sphere in space of n dimensions (n = number of inputs)
node with weight vector closest to input vector is winner
Linear normalisation [ wij = 1] all weights on a plane
12
Geometric interpretation II
Weight vectors move towards input in the hyper space
wij= g * ai/nk - g * wij
Output nodes move towards clusters in inputs
© Rumelhart & Zipser, 1986
13
Stable / unstable
Output nodes move towards clusters in inputs
If input not clustered...
...output nodes will continue moving through input space! © Rumelhart & Zipser, 1986
14
Statistical equivalents
Sarle (1994):Classification = k-means clustering
Kohonen = mapping continuous dimensions onto discrete ones
Statistical techniques usually more efficient...
...because statistical techniques use whole data set
15
Importance of competitive learning Supervised - unsupervised learning
Structure input sets not always given
Natural categories
16
Competitive learning in the brain Lateral inhibition feature of most parts of
the brain
… Implements winner-take-all ?
18
Map formation in the brain
Topographic maps omnipresent in the sensory regions of the brain
retinotopic maps: neurons ordered as the locations of their visual field on the retina
tonotopic maps: neurons ordered according to tone for which they are sensitive
maps in somatosensory cortex: neurons ordered according to body part for which they are sensitive
maps in motor cortex: neurons ordered according to muscles they control
21
Speculations
Map formation ubiquitous (also semantic maps?)
How do maps form? gradients in neurotransmitters pruning
22
Kohonen maps
Teuvo Kohonen first to show how maps can develop
Self Organising Maps (S.O.M.)
Demonstration: the ordering of colours (colours are vectors in a 3-dimensional space of brightness, hue, saturation).
23
Kohonen algorithm
Finding the activity bubble
Updating the weights for the nodes in the active bubble
26
Updating the weights
Move weight vector of winner towards the input vector
Do the same for the active neighbourhood nodes
weight vectors of neigbouring nodes will start resembling each other
27
Simplest implementation
Weight vectors & input patterns all have length 1 (e.i., (wij)
2 = 1 ) Find node whose weight vector has mimimal
distance to the input vector:min. (aj - wij)2
Activate all nodes in neighbourhood radius Nt
Update weights of active nodes by moving weights towards the input vector:
wij = t * ( aj - wij)
wij(t+1) = wij(t) + t * ( aj - wij(t) )
29
Influence of neighbourhood radius
© Kohonen, 1982
Larger neighbourhood size leads to faster learning