0 Voturi pozitive0 Voturi negative

0 (de) vizualizări97 paginiMar 20, 2019

Lesson_3.6_-_Supervised_Learning_Neural_Networks (1).pdf

© © All Rights Reserved

PDF, TXT sau citiți online pe Scribd

© All Rights Reserved

0 (de) vizualizări

Lesson_3.6_-_Supervised_Learning_Neural_Networks (1).pdf

© All Rights Reserved

- Steve Jobs
- The Woman Who Smashed Codes: A True Story of Love, Spies, and the Unlikely Heroine who Outwitted America's Enemies
- NIV, Holy Bible, eBook
- NIV, Holy Bible, eBook, Red Letter Edition
- Cryptonomicon
- Hidden Figures Young Readers' Edition
- Everybody Lies: Big Data, New Data, and What the Internet Can Tell Us About Who We Really Are
- Make Your Mind Up: My Guide to Finding Your Own Style, Life, and Motavation!
- Console Wars: Sega, Nintendo, and the Battle that Defined a Generation
- The Golden Notebook: A Novel
- Alibaba: The House That Jack Ma Built
- Life After Google: The Fall of Big Data and the Rise of the Blockchain Economy
- Hit Refresh: The Quest to Rediscover Microsoft's Soul and Imagine a Better Future for Everyone
- Hit Refresh: The Quest to Rediscover Microsoft's Soul and Imagine a Better Future for Everyone
- The Innovators: How a Group of Hackers, Geniuses, and Geeks Created the Digital Revolution
- The 10X Rule: The Only Difference Between Success and Failure
- Autonomous: A Novel

Sunteți pe pagina 1din 97

Anis ur Rahman

Department of Computing

NUST-SEECS

Islamabad

1 / 97

Prediction by Neural Networks

Roadmap

Artificial Neural Networks

Simple Computing Elements

Feed-Forward Networks

Perceptrons (Single-layer, Feed-Forward Neural Network)

Perceptron Learning Method

Multilayer Feed-Forward Neural Network

Defining a Network Topology

Backpropagation Algorithm

Backpropagation and Interpretability

Discussion

2 / 97

Prediction by Neural Networks

the fundamental functional unit of all nervous system tissue,

including the brain.

1011 neurons in the human brain

Neuron components

1 Soma (cell body). provides the support functions and structure of

the cell, that contains a cell nucleus.

2 Dendrites. consist of more branching fibers which receive signal

from other nerve cells

3 Axon.

a branching fiber which carries signals away from the neuron that

connect to the dendrites and cell bodies of other neurons.

In reality, the length of the axon should be about 100 times the

diameter of the cell body.

4 Synapse. The connecting junction between axon and dendrites.

3 / 97

Prediction by Neural Networks

4 / 97

Prediction by Neural Networks

cell body

2 When a potential of cell body reaches some limit, neuron “fires”,

electrical signal (action potential) sent down axon

3 Axon propagates signal to other neurons, downstream

5 / 97

Prediction by Neural Networks

Excitatory synapse. increasing potential

Inhibitory synapse. decreasing potential

Synaptic plasticity. controls how effectively two neurons

communicate with each other.

Migration of neurons

Neurons also form new connections with other neurons.

Sometimes entire collections of neurons can migrate from one

place to another.

These mechanisms are thought to form the basis for learning in

the brain.

consciousness.

6 / 97

Prediction by Neural Networks

Parallelism. all the neurons and synapses are active

simultaneously, whereas most current computers have only one or

at most a few CPUs.

More fault-tolerant. A hardware error that flips a singlebit can

doom an entire computation, but brain cells die all the time with no

ill effect to the overall functioning of the brain.

Inductive algorithm. To be trained using an inductive learning

algorithm

7 / 97

Prediction by Neural Networks

computational analogues of neurons

Other names:

connectionist learning,

parallel distributed processing,

neural computation,

adaptive networks, and

collective computation

8 / 97

Prediction by Neural Networks

Components:

1 Units

A neural network is composed of a number of nodes, or units

Metaphor for nerve cell body

2 Links

Units connected by links

Links represent synaptic connections from one unit to another

3 Weight

Each link has a numeric weight

9 / 97

Prediction by Neural Networks

An example of ANN

10 / 97

Prediction by Neural Networks

Long-term memory

Weights are the primary means of long-term storage in neural

networks

Learning method

Learning usually takes place by adjusting the weights

Input and Output Units

Some of the units are connected to the external environment, and

can be designated as input units or output units

11 / 97

Prediction by Neural Networks

Components of a Unit

1 a set of input links from other units,

2 a set of output links to other units,

3 a current activation level, and

4 a means of computing the activation level at the next step in time,

given its inputs and weights.

The idea is that each unit does a local computation based on inputs

from its neighbors, but without the need for any global control over

the set of units as a whole.

12 / 97

Prediction by Neural Networks

Soma / Cell body ⇔ Neuron / Node / Unit

Dendrite ⇔ Input links

Axon ⇔ Output links

Synapse ⇔ Weight

13 / 97

Prediction by Neural Networks

supervised learning, and

unsupervised learning

For supervised learning neural networks can be used for both

classification. to predict the class label of a given example, and

prediction. to predict a continuous-valued output

In this lesson we want to discuss about application of neural networks

for supervised learning.

14 / 97

Prediction by Neural Networks

how many units are to be used

what kind of units are appropriate

how the units are to be connected to form a network

2 Then

initializes the weights of the network, and

trains the weights using a learning algorithm applied to a set of

training examples for the task

The use of examples also implies that one must decide how to encode

the examples in terms of inputs and outputs of the network.

15 / 97

Prediction by Neural Networks

1 Receives n-inputs

2 Multiplies each input by its weight

3 Applies activation function to the sum of results

4 Outputs result

16 / 97

Prediction by Neural Networks

1 Linear component

input function, ini , that computes the weighted sum of the unit’s

input values

2 Non-linear component

activation function, g, that transforms the weighted sum into the

final value that serves as the unit’s activation value, ai

Usually, all units in a network use the same activation function

17 / 97

Prediction by Neural Networks

A typical unit

18 / 97

Prediction by Neural Networks

Total weighted input

X

ini = Wj ,i aj

j

The input values is called aj

19 / 97

Prediction by Neural Networks

Processing:

3(0.3) + 1(−0.1) + 0(2.1) + −1.1(−2)

= 0.9 + (−0.1) + 2.2

=3

20 / 97

Prediction by Neural Networks

X

ai = g(ini ) = g Wj ,i aj

j

1 Step function

2 Sign function

3 Sigmoid function

21 / 97

Prediction by Neural Networks

(

1 if x ≥ t

stept (x) =

0 if x < t

(

+1 if x ≥ 0

sign(x) =

−1 if x < 0

1

sigmoid (x) =

1 + e −x

22 / 97

Prediction by Neural Networks

Step Function

outputs a 1 when the input is greater than its threshold, and

outputs a 0 otherwise

The biological motivation is that a 1 represents the firing of a

pulse down the axon, and a 0 represents no firing.

The threshold represents the minimum total weighted input

necessary to cause the neuron to fire.

23 / 97

Prediction by Neural Networks

Let t = 4

Step4 (3) = 0

24 / 97

Prediction by Neural Networks

Step Function

extra input weight

Because it need only worry about adjusting weights, rather than

adjusting both weights and thresholds

Thus, instead of having a threshold t for each unit, we add an extra

input whose activation

Xn Xn

ai = stept Wj ,i aj = step0 Wj ,i aj

j =1 j =0

where W0,i = t and a0 = −1

25 / 97

Prediction by Neural Networks

Step Function

The Figure shows how the Boolean functions AND, OR, and NOT

can be represented by units with a step function and suitable

weights and thresholds.

This is important because it means we can use these units to build

a network to compute any Boolean function of the inputs.

26 / 97

Prediction by Neural Networks

Sigmoid Function

1

f (x) =

1 + e −σ x

where σ : the steepness parameter

27 / 97

Prediction by Neural Networks

Sigmoid Function

1

f (x) =

1 + e −σ x

1

f (3) = ≈ 0.95

1 + e −x

28 / 97

Prediction by Neural Networks

Sigmoid Function

29 / 97

Prediction by Neural Networks

Another Example

Two inputs, one output, one ‘hidden’ unit

Input: (3, 1)

1

f (x) =

1 + e −x

30 / 97

Prediction by Neural Networks

Computing:

1 Start at leftmost layer

2 Compute activations based on inputs

3 Then work from left to right, using computed activations as inputs to

next layer

Example solution:

Activation of hidden unit

f (0.5(3) + −0.5(1)) = f (1.5 − 0.5) = f (1) = 0.731

Output activation

f (0.731(0.75)) = f (0.548) = 0.634

31 / 97

Prediction by Neural Networks

Feed-Forward Networks

Unidirectional links

Directed acyclic (no cycles) graph (DAG)

No links between units in the same layer

No links backward to a previous layer

No links that skip a layer

Uniformly processing from input units to output units

32 / 97

Prediction by Neural Networks

Feed-forward Networks

and one output node.

33 / 97

Prediction by Neural Networks

Feed-forward Networks

Types of units

Input units. the activation value of each of these units is

determined by the environment

Output units. at the right-hand end of the network units

Hidden units. they have no direct connection to the outside

world

Because the input units (square nodes) simply serve to pass activation

to the next layer, they are not counted

34 / 97

Prediction by Neural Networks

Feed-forward Networks

1 Perceptrons

No hidden units

This makes the learning problem much simpler, but

it means that perceptrons are very limited in what they can

represent

2 Multilayer networks

one or more hidden units

35 / 97

Prediction by Neural Networks

Feed-forward Networks

activation functions g

The functions have a specific parameterized structure

The weights chosen for the network determine which of these

functions is actually represented

For example, the network calculates the following function:

a5 = g(W3,5 a3 + W4,5 a4 )

= g(W3,5 (W1,3 a1 + W2,3 a2 ) + W4,5 (W1,4 a1 + W2,4 a2 ))

36 / 97

Prediction by Neural Networks

network represents a complex nonlinear function

If you think of the weights as parameters or coefficients of this

function, then learning just becomes:

a process of tuning the parameters to fit the data in the training

set—a process that statisticians call nonlinear regression

37 / 97

Prediction by Neural Networks

incapable of representation

2 Too big network

not generalized well

Overfitting when there are too many parameters

38 / 97

Prediction by Neural Networks

Perceptrons

Perceptrons

Single-layer feed-forward network

first studied in the late 1950s

Types of Perceptrons

Single-output perceptron. with a single output unit

Multi-output perceptron. with several output units

39 / 97

Prediction by Neural Networks

Perceptrons

Each weight only affects one of the outputs

40 / 97

Prediction by Neural Networks

Perceptrons

Activation of output unit:

X

O = Step0 Wj Ij = Step0 (W · I )

j

where

Wj : weight from input unit j

Ij : activation of input unit j

the step function, with I0 = −1.

41 / 97

Prediction by Neural Networks

Perceptrons

represent

The problem is that any input Ij can only influence the final output

in one direction, no matter what the other input values are

Suppose that this vector has aj = 0 and that the vector produces a 0

as output. Furthermore, suppose that when aj is replaced with 1,

the output changes to 1. This implies that Wj must be positive

It also implies that there can be no input vector b for which the

output is 1 when bj = 0, but the output is 0 when bj is replaced with

1

42 / 97

Prediction by Neural Networks

Perceptrons

Three different Boolean functions of two inputs, the AND, OR, and

XOR functions

Black dots indicate a point in the input space where the value of the

function is 1, and white dots indicate a point where the value is 0.

separates all the white dots from the black dots

Such functions are called linearly separable

Thus, a perceptron can represent AND and OR, but not XOR (if Il # I2 ).

43 / 97

Prediction by Neural Networks

Perceptrons

functions follows directly from equation:

X

O = Step0 Wj Ij = Step0 (W · I )

j

This means that the entire input space is divided in two along a

boundary defined by W · I = 0,

that is, a plane in the input space with coefficients given by the

weights.

44 / 97

Prediction by Neural Networks

Perceptrons

one possible separating “plane” is the dotted line defined by the

equation

I1 = −I2 + 1.5 or I1 + I2 = 1.5

The region above the line, where the output is 1, is therefore given

by

−1.5 + I1 + I2 > 0

45 / 97

Prediction by Neural Networks

the range [−0.5, 0.5]

2 The network is then updated to try to make it consistent with the

training examples (instances)

This is done by making small adjustments in the weights to reduce

the difference between the observed and predicted values

3 The algorithm is the need to repeat the update phase several

times for each example in order to achieve convergence

Each epoch involves updating all the weights for all the examples

46 / 97

Prediction by Neural Networks

47 / 97

Prediction by Neural Networks

If the predicted output for the single output unit is O , and the correct

output should be T , then the error is given by

Err = T − O

2 If the Err is negative, we need to decrease O

1 If Ij is positive, an increase in Wj will tend to increase O

2 If Ij is negative, an increase in Wj will tend to decrease O

48 / 97

Prediction by Neural Networks

Wj ← Wj + α ∗ Ij ∗ Err

where α : is the learning rate

Frank Rosenblatt

Remarks

Rosenblatt proved that a learning system using the perceptron learning

rule will converge to a set of weights that correctly represents the

examples, as long as the examples represent a linearly separable

function.

49 / 97

Prediction by Neural Networks

α: Learning rate

T : Target or correct output

O : Net (summed, weighted) input to output unit

Ij : j th input value

50 / 97

Prediction by Neural Networks

Example

W = (W1 , W2 , W3 )

Initially: W = (.5 .2 .4)

Let α = 0.5

Apply delta rule

1 000 0

2 111 1

3 100 1

4 001 1

51 / 97

Prediction by Neural Networks

output (T) output (O) weights updates

1 (0 0 0) 0 0 (.5 .2 .4)

2 (1 1 1) 1

3 (1 0 0) 1

4 (0 0 1) 1

52 / 97

Prediction by Neural Networks

output (T) output (O) weights updates

1 (0 0 0) 0 0 (.5 .2 .4) W1 : 0.5(0 − 0)0

W2 : 0.5(0 − 0)0

W3 : 0.5(0 − 0)0

53 / 97

Prediction by Neural Networks

output (T) output (O) weights updates

1 (0 0 0) 0 0 (.5 .2 .4) (0 0 0)

2 (1 1 1) 1 (.5 .2 .4)

3 (1 0 0) 1

4 (0 0 1) 1

54 / 97

Prediction by Neural Networks

output (T) output (O) weights updates

1 (0 0 0) 0 0 (.5 .2 .4) (0 0 0)

2 (1 1 1) 1 1.1 (.5 .2 .4) (-.05 -.05 -.05)

3 (1 0 0) 1 .45 (.45 .15 .35) (.275 0 0)

4 (0 0 1) 1 .35 (.725 .15 .35) (0 0 .325)

55 / 97

Prediction by Neural Networks

After 18 epochs

Weights

W1 = 0.990735

W2 = −0.970018005

W3 = 0.98147

1 000 0

2 111 1

3 100 1

4 001 1

56 / 97

Prediction by Neural Networks

Example

Actual Outputs

1 000 0 0

2 111 1 1.002187

3 100 1 0.990735

4 001 1 0.98147

57 / 97

Prediction by Neural Networks

Examples in ANN

for neural networks and those used for other attribute-based

methods such as decision trees

In a neural network, all inputs are real numbers in some fixed

range, whereas decision trees allow for multivalued attributes

with a discrete set of values

For example, an attribute may have values None, Some, and Full

58 / 97

Prediction by Neural Networks

1 Local encoding

Use a single input unit, and pick an appropriate number of distinct

values to correspond to the discrete attribute values

For example, we can use None = 0.0, Some = 0.5, and Full = 1.0

2 Distributed encoding

Use one input unit for each value of the attribute, turning on the

unit that corresponds to the correct value

59 / 97

Prediction by Neural Networks

layers includes:

1 an input layer,

2 one or more hidden layers, and

3 an output layer

A two-layer neural network has a hidden layer and an output layer

The input layer is not counted because it serves only to pass the

input values to the next layer

A network containing two hidden layers is called a three-layer

neural network, and so on.

60 / 97

Prediction by Neural Networks

How many input units when ten attributes describe each example?

We will need ten input units

How many hidden units are needed?

The problem of choosing the right number of hidden units in

advance is still not well-understood

We use a network with four hidden units

61 / 97

Prediction by Neural Networks

62 / 97

Prediction by Neural Networks

Inputs are fed simultaneously into the units making up the input

layer

The inputs to the network correspond to the attributes measured

for each training example

They are then weighted and fed simultaneously to a hidden

layer

The number of hidden layers is arbitrary, although in practice,

usually only one is used

The weighted outputs of the last hidden layer are input to units

making up the output layer, which sends out the network’s

prediction

63 / 97

Prediction by Neural Networks

back to an input unit or to an output unit of a previous layer

From a statistical point of view, networks perform nonlinear

regression

Given enough hidden units and enough training samples, they can

closely approximate any function

64 / 97

Prediction by Neural Networks

Learning method

1 Example inputs are presented to the network and the network

computes an output vector that matches the target

2 If there is an error (a difference between the output and target),

then the weights are adjusted to reduce this error

The trick is to assess the blame for an error and divide it among

the contributing weights

In perceptrons, this is easy, because there is only one weight

between each input and the output

But in multilayer networks, there are many weights connecting each

input to an output, and each of these weights contributes to more

than one output

65 / 97

Prediction by Neural Networks

1 the number of units in the input layer

2 the number of hidden layers (if > 1)

3 the number of units in each hidden layer

4 the number of units in the output layer

training examples to [0.0—1.0] will help speed up the learning phase.

66 / 97

Prediction by Neural Networks

Input units

Discrete-valued attributes may be encoded such that there is one input

unit per domain value.

namely {a0 , a1 , a2 }, then

we may assign three input units to represent A, that is, say, I0 , I1 , I2

as input units

Each unit is initialized to 0

Then

I0 is set to 1, If A = a1

I1 is set to 1, If A = a2

I2 is set to 1, If A = a3

67 / 97

Prediction by Neural Networks

Output unit

1 For classification, one output unit may be used to represent two

classes (where the value 1 represents one class, and the value 0

represents the other)

2 If there are more than two classes, then one output unit per

class is used

68 / 97

Prediction by Neural Networks

There are no clear rules as to the “best” number of hidden layer

units

accuracy of the resulting trained network

repeat the training process with

a different network topology or

a different set of initial weights

69 / 97

Prediction by Neural Networks

2 Hill-climbing search. modifying an existing network structure

Start with a big network: optimal brain damage algorithm

3 Pruning. Removing weights from fully connected model

Start with a small network: tiling algorithm

4 Start with single unit and add subsequent units

5 Cross-validation techniques. are useful for deciding when we

have found a network of the right size

70 / 97

Prediction by Neural Networks

Backpropagation

multilayer feed-forward neural network

It is the most popular method for learning in multilayer networks

Backpropagation iteratively process a set of training examples &

compare the network’s prediction with the actual known target

value

The target value may be the known class label of the training

example (for classification problems) or a continuous value (for

prediction problems)

71 / 97

Prediction by Neural Networks

Backpropagation

the mean squared error between the network’s prediction and the

actual target value

Modifications are made in the “backwards” direction

from the output layer, through each hidden layer down to the first

hidden layer, hence “backpropagation”

Although it is not guaranteed, in general the weights will

eventually converge, and the learning process stops

72 / 97

Prediction by Neural Networks

Backpropagation

1 Initialize the weights

Initialize weights to small random and biases in the network

2 Propagate the inputs forward

by applying activation function

3 Backpropagate the error

by updating weights and biases

4 Terminating condition

when error is very small, etc

73 / 97

Prediction by Neural Networks

Backpropagation Algorithm

1 Input:

D , a data set consisting of the training examples and their

associated target values

l , the learning rate

network, a multilayer feed-forward network

2 Output:

A trained neural network

74 / 97

Prediction by Neural Networks

The weights in the network are initialized to small random numbers

e.g., ranging from -1.0 to 1.0 or -0.5 to 0.5

Each unit has a bias associated with it

The biases are similarly initialized to small random numbers

Each training example is processed by the steps 2 to 8.

75 / 97

Prediction by Neural Networks

The training example is fed to the input layer of the network

The inputs pass through the input units, unchanged

For an input unit, j ,

its input value, Ij

its output, Oj , is equal to its input value, Ij .

76 / 97

Prediction by Neural Networks

3) Compute the net input of each unit in the hidden and output layers

The net input to a unit in the hidden or output layers is computed

as a linear combination of its inputs

Given a unit j in a hidden or output layer, the net input, Ij , to unit j is

X

Ij = wij Oi + θj

i

where wij is the weight of the connection from unit i in the previous

layer to unit j

Oi is the output of unit i from the previous layer

θj is the bias of the unit

77 / 97

Prediction by Neural Networks

78 / 97

Prediction by Neural Networks

4) Compute the output of each unit j in the hidden and output layers

The output of each unit is calculating by applying an activation

function to its net input

The logistic, or sigmoid, function is used

Given the net input Ij to unit j , then Oj , the output of unit j , is

computed as:

1

Oj =

1 + e −Ij

79 / 97

Prediction by Neural Networks

For a unit j in the output layer, the error Errj is computed by

Errj = Oj (1 − Oj )(Tj − Oj )

Tj is the known target value of the given training example

Note that Oj (1 − Oj ) is the derivative of the logistic function

80 / 97

Prediction by Neural Networks

6) Compute the error for each unit j in the hidden layers, from the last

to the first hidden layer

The error of a hidden layer unit j is

X

Errj = Oj (1 − Oj ) Errk wjk

k

next higher layer, and

Errk is the error of unit k

81 / 97

Prediction by Neural Networks

Weights are updated by the following equations

wij = wij + ∆wij

∆wij = (l )Errj Oi

The variable l is the learning rate, a constant typically having a

value between 0.0 and 1.0

82 / 97

Prediction by Neural Networks

Learning rate

Backpropagation learns using a method of gradient descent

The learning rate helps avoid getting stuck at a local minimum in

decision space (i.e., where the weights appear to converge, but are

not the optimum solution) and encourages finding the global

minimum

If the learning rate is too small, then learning will occur at a very

slow pace

If the learning rate is too large, then oscillation between

inadequate solutions may occur

A rule to set the learning rate to 1t , where t is the number of

iterations through the training set so far

83 / 97

Prediction by Neural Networks

Biases are updated by the following equations below:

θj = θj + ∆θj

∆θj = (l )Errj

There are two strategies for updating the weights and biases

84 / 97

Prediction by Neural Networks

Updating strategies:

1 Case updating

updating the weights and biases after the presentation of each

example.

case updating is more common because it tends to yield more

accurate result

2 Epoch updating

The weight and bias increments could be accumulated in variables,

so that the weights and biases are updated after all of the examples

in the training set have been presented.

One iteration through the training set is an epoch.

85 / 97

Prediction by Neural Networks

Terminating Condition

After finishing the processed for all training examples, we must

evaluate the stopping condition

Stopping condition: Training stops when

All ∆wij in the previous epoch were so small as to be below some

specified threshold, or

The percentage of examples misclassified in the previous epoch is

below some threshold, or

A prespecified number of epochs has expired.

In practice, several hundreds of thousands of epochs may be

required before the weights will converge.

If stopping condition was not true steps 2 to 8 should repeat for all

training examples

86 / 97

Prediction by Neural Networks

Efficiency of Backpropagation

the network.

However, in the worst-case scenario, the number of epochs can be

exponential in n, the number of inputs.

In practice, the time required for the networks to converge is

highly variable.

A number of techniques exist that help speed up the training time.

Metaheuristic algorithms such as simulated annealing algorithm

can be used, which also ensures convergence to a global optimum.

87 / 97

Prediction by Neural Networks

Example

88 / 97

Prediction by Neural Networks

Example

the first training example, X

Let the learning rate be 0.9

The initial weight and bias values of the network are given in the

Table, along with the first training example, X = (1, 0, 1), whose

class label is 1

1 0 1 0.2 -0.3 0.4 0.1 -0.5 0.2 -0.3 -0.2 -0.4 0.2 0.1

89 / 97

Prediction by Neural Networks

Example

The net input and output calculations:

X

Ij = wij Oi + θj

i

1

Oj =

1 + e −Ij

5 −0.3 + 0 + 0.2 + 0.2 = 0.1 1/(1 + e −0.1 ) = 0.525

6 (−0.3)(0.332) − (0.2)(0.525) + 0.1 = −0.105 1/(1 + e 0.105 ) = 0.474

90 / 97

Prediction by Neural Networks

Example

The output layer

Errj = Oj (1 − Oj )(Tj − Oj )

X

Errj = Oj (1 − Oj ) Errk wjk

k

Unit j Errj

6 (0.474)(1 - 0.474)(1 - 0.474) = 0.1311

5 (0.525)(1 - 0.525)(0.1311)(-0.2) = -0.0065

4 (0.332)(1 - 0.332)(0.1311)(-0.3) = -0.0087

91 / 97

Prediction by Neural Networks

Example

θj = θj + ∆θj

where∆wij = (l )Errj Oi and ∆θj = (l )Errj

w46 -0.3 + (0.9)( 0.1311)(0.332) = -0.261

w56 -0.2 + (0.9)( 0.1311)(0.525) = -0.138

w14 0.2 + (0.9)(-0.0087)(1) = 0.192

w15 -0.3 + (0.9)(-0.0065)(1) = -0.306

w24 0.4 + (0.9)(-0.0087)(0) = 0.4

w25 0.1 + (0.9)(-0.0065)(0) = 0.1

w34 -0.5 + (0.9)(-0.0087)(1) = -0.508

w35 0.2 + (0.9)(-0.0065)(1) = 0.194

θ6 0.1 + (0.9)( 0.1311) = 0.218

θ5 0.2 + (0.9)(-0.0087) = 0.194

θ4 -0.4 + (0.9)(-0.0065) = -0.408

92 / 97

Prediction by Neural Networks

Example

algorithm have been proposed for classification in neural networks.

These may involve:

the dynamic adjustment of the network topology and of the

learning rate

New parameters

The use of different error functions

93 / 97

Prediction by Neural Networks

A major disadvantage of neural networks lies in their knowledge

representation

Acquired knowledge in the form of a network of units connected by

weighted links is difficult for humans to interpret

This factor has motivated research in extracting the knowledge

embedded in trained neural networks and in representing that

knowledge symbolically

94 / 97

Prediction by Neural Networks

Discussion

High tolerance to noisy data

It can be used when you may have little knowledge of the

relationships between attributes and classes

Well-suited for continuous-valued inputs and outputs

Successful on a wide array of real-world data

Algorithms are inherently parallel

Techniques have recently been developed for the extraction of

rules from trained neural networks

95 / 97

Prediction by Neural Networks

Research Areas

e.g. by genetic algorithms

Increasing learning speed (efficiency)

e.g. by simulated annealing

Increasing accuracy (effectiveness)

Extracting rules from networks

96 / 97

Prediction by Neural Networks

References

Inc. (2006). (Chapter 6)

S. J. Russell and P. Norvig, Artificial Intelligence, A Modern

Approach, Prentice Hall,1995. (Chapter 19)

97 / 97

- Neuroplasticity what is itÎncărcat deStella Papadopoulou
- Artificial Neural Networks_MiniProjectÎncărcat deamithbalu
- GeopolymerÎncărcat deCozy King Leishangthem
- das2014.pdfÎncărcat degoutham
- Comparative Analysis of Power System Stabilizer using Artificial Intelligence TechniquesÎncărcat deInternational Journal for Scientific Research and Development - IJSRD
- Sistema NerviosoÎncărcat deNataly Quintero
- Data-Mining and Knowledge Discovery, Neural Networks InÎncărcat deElena Galeote Lopez
- 1-s2.0-S0167268108001959-mainÎncărcat deSam Keays
- report.docxÎncărcat deAnimesh Biswal
- Generalized Regression Neural Network - A better Predictive Tool for Estimating PVT Properties of Niger Delta Crude OilÎncărcat deBasil Ogbunude
- Combining Genetic Algorithm and Neural NetworksÎncărcat deVivek Bhadouria
- Incremental Learning Approach for Enhancing the Performance of Multi-Layer Perceptron for Determining the Stock TrendÎncărcat deDr Vishal Jain
- ANN CASE STUDYÎncărcat desuryays
- WorkloadÎncărcat deMagesh Govindarajan
- Algorithms 02 00973Încărcat debaruah_renika
- mlmlmlmlml.pdfÎncărcat deCalvin Low
- IIJCS-2014-02-28-019Încărcat deAnonymous vQrJlEN
- video_6Încărcat dehawxfly
- Vision the Journal of Business Perspective 2011 MuVikhopadhyay 153 62Încărcat deFeroz Ikbal
- A Competency Based Management System for Sustainable Development by InnovativeÎncărcat desamas7480
- ANN 2009 (2)Încărcat desoran-civel
- Breast Cancer1Încărcat deabhishek
- Prediction of Coal Grindability Based on Petrography, Proximate and Ultimate Analysis Using Multiple Regression and Artificial Neural Network ModelsÎncărcat deRizal Ahmad Mubarok
- nn35Încărcat devenkatesh_pabolu
- 1Încărcat demariah.hafeez4474
- EE6224-2009-1Încărcat dephatct
- paper main.docxÎncărcat deranvijay
- Teoría de Marín-PadillaÎncărcat deDaniela Güiza Argüello
- The Nerve System 1Încărcat deFarah Nabila Luthfiyya
- 10.1.1.193Încărcat deDevendra Kumar Dubey

- Lesson 4.1 - Unsupervised Learning Partitioning MethodsÎncărcat deTayyaba Faisal
- Lesson_2.1_-_Know_Your_Data.pdfÎncărcat deTayyaba Faisal
- Lesson_4.1_-_Unsupervised_Learning_Partitioning_Methods.pdfÎncărcat deTayyaba Faisal
- Lesson 3.6 - Supervised Learning Neural NetworksÎncărcat deTayyaba Faisal
- Lesson_3.2_-_Supervised_Learning_Evaluation.pdfÎncărcat deTayyaba Faisal
- Lesson 2.2 - Frequent Pattern AnalysisÎncărcat deTayyaba Faisal
- Lesson 3.3 - Supervised Learning Rule Based ClassificationÎncărcat deTayyaba Faisal
- Lesson 3.1 - Supervised Learning Decision TreesÎncărcat deTayyaba Faisal
- Lesson 3.2 - Supervised Learning EvaluationÎncărcat deTayyaba Faisal

- Steps Cloning Database With RMANÎncărcat deMabu Dba
- Borang Bekalan Data Swasta - ForM DI.2Încărcat deHushes Cheong
- Efficycle 2018 Information Document for Motor KitÎncărcat deSonu Raghuwanshi
- LCCI002 Certificate in ICT Practical SkillsÎncărcat deAnonymous I6KIRj2Wv
- Chapter 12 - Network Design and Facility LocationÎncărcat deArman
- great-lakes-pgp-babi-brochure.pdfÎncărcat deVashirAhmad
- L293D Motor Driver ICÎncărcat deajayrana424
- 27 - Elementos Gráficos e Multimídia JavaFXÎncărcat deTaciano Canassa
- AlmaCenaMiento en apps para Windows StoreÎncărcat deiberhack
- Object Oriented Modeling & Design [10CS71]-NOTES.docxÎncărcat deDenny Arshan
- Instruction of ZTE V9820 SD Card UpgradeÎncărcat deAgung Priadi
- Wii Donkey Kong Country ReturnsÎncărcat deRoGdiN
- Ctv Old IptvÎncărcat deMuhamad Zaki
- BITCOIN FINANCIAL REGULATION: SECURITIES, DERIVATIVES, PREDICTION MARKETS, AND GAMBLINGÎncărcat deSamuel Hammond
- Arindam Mukhopadhyay (Alu) ProfileÎncărcat dechatuusumitava
- Assymptotic NotationsÎncărcat deAhmed Ajmal
- Intelligent Gsm Care for Bus Finding SystemÎncărcat deManasa Chikkala
- tlbmimoÎncărcat deShriram Swaminathan
- CBSE class 12 physics investigatory project on logic gatesÎncărcat deRohan Patel
- Loyalty Management November 2011Încărcat deLoyalty360
- Sensor Course WorkÎncărcat deYuan Gao
- Operations Cheat SheetÎncărcat demathi
- Samsung 731N Menu Keeps Popping UpÎncărcat detarkot
- Module_4 Grouping and Aggregating Dataod1Încărcat decviga
- Ecwid TutorialÎncărcat deYourAdsStrategist
- Hybrid Inverse Airfoil Design Method for Complex Three-Dimensional Lifting SurfacesÎncărcat dekrishnamoorthy_krish
- addressing modesÎncărcat deAruna Konduri
- Root CauseAnalysisÎncărcat deHoang
- Topic 3 - User Centered Design - RAÎncărcat deRay Sunny
- LRS-350-SPEC.PDFÎncărcat deDevdatta Mhaiskar

## Mult mai mult decât documente.

Descoperiți tot ce are Scribd de oferit, inclusiv cărți și cărți audio de la editori majori.

Anulați oricând.