Documente Academic
Documente Profesional
Documente Cultură
2.3 Gradient descent algorithm for training Fig. 4 Description of BPTT for Weight Update
neural networks
Error backpropagation employs gradient descent 2.4 Genetic algorithms for training neural
learning and is the most popular algorithm used for training networks
neural networks. The goal of gradient descent learning is to
minimize the sum of squared errors by propagating error Genetic algorithms provide a learning method motivated
signals backward through the network architecture upon the by biological evolution [20]. They have been successfully
presentation of training samples from the training set. These applied to neural network weight updates and to network
error signals are used to calculate the weight updates which topology optimization [21]. In recent years, such hybrid
represent the knowledge learnt from training. A limitation of approaches to neural networks training have gained popularity
gradient descent learning is their tendency of getting trapped and have been applied to real-world problems such as job
in a local minimum during training resulting in poor training scheduling [22], forecasting [23] and robotics control [24].
and generalization performance. The general idea in using genetic algorithms for training
Backpropagation is used for training feedforward networks neural networks is to encode weights as chromosomes in a
while backpropagation-through-time (BPTT) is employed for population. The task of genetic algorithms then is to find
training recurrent neural networks. BPTT is the spatio- optimal sets of weights that best represent the knowledge after
temporal extension of the backpropagation algorithm [18]. being presented with the training data in the network. The
The general idea behind BPTT is to unfold the recurrent fitness function is thus the sum of squared errors returned by
the network after being presented with the weights encoded in
chromosomes. Genetic algorithms find the optimal set of
weights in a network topology which minimizes the error
function. To evaluate the fitness function, each weight
encoded in the chromosome is assigned to the respective
weight links of the network. The training set of examples is
then presented to the network which propagates the input
signals forward and the sum of squared errors is calculated.
In this way, genetic algorithms attempt to find a set of weights
which minimizes the error function of the network. Unlike
learning with gradient descent, genetic algorithms can help
neural networks to escape from the local minima in weight
space. Fig 5 shows a high level description of genetic
algorithms for training RNNs.
____________________________