Documente Academic
Documente Profesional
Documente Cultură
Primary reinforcement
Environment Critics
Heuristic
reinforcement
Learning
Systems
Terminology:
Training set: The ensemble of “inputs” used to train
the system. For a supervised network. It is the
ensemble of “input-desired” response pairs used to
train the system.
Validation set: The ensemble of samples that will be
used to validate the parameters used in the training
(not to be confused with the test set which assesses
the performance of the classifier).
Test set: The ensemble of “input-desired” response
data used to verify the performance of a trained
system. This data is not used for training.
Training epoch: one cycle through the set of training
patterns.
Generalization: The ability of a NN to produce
reasonable responses to input patterns that are
similar, but not identical, to training patterns.
Terminology:
Asynchronous: process in which weights or
activations are updated one at a time, rather than all
being updated simultaneously.
Synchronous updates: All weights are adjusted at the
same time.
Inhibitory connection: connection link between two
neurons such that a signal sent over this link will
reduce the activation of the neuron that receives the
signal . This may result from the connection having a
negative weight, or from the signal received being
used to reduce the activation of a neuron by scaling
the net input the neuron receives from other neurons.
Activation: a node’s level of activity; the result of
applying the activation function to the net input to the
node. Typically this is also the value the node
transmits.
Review:
Vectors- Overview
Vectors- A Brief review
•Compute distances
K nearest neighbors
New item
•Compute distances
•Pick K best distances
K nearest neighbors
New item
•Compute distances
•Pick K best distances
•Assign class to new
example
Example: image search
Query image
D. Hebb
LR3: Hebbian Learning
“When an axon of cell A is near enough to excite a cell B and
repeatedly or persistently takes place in firing it, some growth
process or metabolic change takes place in one or both cells such
that A’s efficiency, as one of the cells firing B, is increased” (Hebb,
1949)
In other words:
1. If two neurons on either side of a synapse (connection) are activated
simultaneously (i.e. synchronously), then the strength of that synapse is
selectively increased.
This rule is often supplemented by:
2. If two neurons on either side of a synapse are activated
asynchronously, then that synapse is selectively weakened or
eliminated. so that chance coincidences do not build up connection
strengths.
LR3: Hebbian Learning
A purely feed forward, unsupervised learning
The learning signal is equal to the neuron’s output
Lateral
connection ( )
is inhibitory
- lateral inhibition
If the net can learn a weight vector configuration like this,
without being told explicitly of the existence of clusters at
the input, then it is said to undergo a process of self-
organised or unsupervised learning. This is to be contrasted
with nets which were trained with the delta rule for e.g.
where a target vector or output had to be supplied.
LR4: Competitive Learning…
In order to achieve this goal, the weight vectors must be
rotated around the sphere so that they line up with the
training set.
The first thing to notice is that this may be achieved in a
gradual and efficient way by moving the weight vector
which is closest (in an angular sense) to the current input
vector towards that vector slightly.
The node k with the closest vector is that which gives the
greatest input excitation v=w.x since this is just the dot
product of the weight and input vectors. As shown below,
the weight vector of node k may be aligned more closely
with the input if a change is made according to
∆ wmj = α (x j − wmj )
LR4: Winner-Take-All learning..
L ( w ) = log ∏ P( X α
xα ∈ℑ
= xα )
= ∑ log P ( X α
xα ∈ℑ
= xα )
−
Let ρ kj denote the correlation between the states of
neurons j and k with network in free-running condition
ρ kj− = ∑ ∑ p(X
x α ∈ℑ x
= x ) xk x j
Note: DON’T PANIC. Boltzmann machine will be presented in details in future lectures.
End of Learning Rules (LR)
Network complexity
No formal methods exist for determining
network architecture. For e.g. the number of
layers in a feed forward network, the number
of nodes in each layer…
The next lectures will focus on specific
networks.
Suggested Reading.
S. Haykin, “Neural Networks”, Prentice-Hall, 1999,
chapter 2, and section 11.7, chapter 11 (for Boltzmann
learning).
L. Fausett, “Fundamentals of Neural Networks”,
Prentice-Hall, 1994, Chapter 2, and Section 7.2.2. of
chapter 7 (for Boltzmann machine).
R.P. Lippmann, “An Introduction to Computing with
Neural Nets”, IEEE Magazine on Acoustics, Signal and
Speech Processing, April 1987: 4-22.
B. Widrow, “Generalization and Information Storage in
Networks of Adaline “neurons”, Self-Organizing
Systems, 1962, ed. MC. Jovitz, G.T. Jacobi, G.
Goldstein, Spartan Books, 435-461
References:
In addition to the references of the previous slide, the
following references were also used to prepare these
lecture notes.