Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response (...

69
Supervised Learning

Transcript of Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response (...

Page 1: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Supervised Learning

Page 2: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Teacher response: Emulation.

Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ).Aim: To reduce the error.

It’s Closed Feedback System.

Suppose the error, Error, is on the surface, due to teacher response, we have to bring it down, which is called minimising error, Point is called Local Minimum, or Global Minimum.

Page 3: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Reinforcement Learning.

Page 4: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

There is no teacher.

Converts Primary reinforcement to heuristic reinforcement.

There can be delay for primary reinforcement, as the inputs have to be analysed, which is called credit assignement problem.

Example Character Recognition.

Page 5: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Unsupervised Learning

Page 6: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Error Correction Learning.

Desired Output – Actual Output

Page 7: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Consider Multi Layer Network.

Page 8: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Multi Layer Network, showing output.

Page 9: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Cost Function or Index of Performance.

Widrow Hoff Rule ( Delta Rule )

New Weights

New Weights ( after Unit Delay )

z (power ) – 1, is called unit delay operator.n is disrete time.

Page 10: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Memory Based Learning

Past experinces are stores, to find the relatiion between Input and desired output.

Consider,

Consider

Page 11: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

K Nearest Neighbour

Page 12: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Hebb’s Associative Rules.

Page 13: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Hebb’s Synapse

Page 14: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Four Characteristics of Hebbian Synapse.

Page 15: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 16: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Hebb’s Model

Hebb’s Hypothesis:

Covariance Hypothesis

Increase in inputs, presynapsis, increases outputs ( postsynapsis), leads to saturation.( Activity Product Rule)

Here thresholds are used on inputs and outputs.

Page 17: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 18: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 19: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 20: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 21: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Output Function.

Summation of Weights.

Change in Weights.

Page 22: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Xj is input and xk is output

T is pseudotempearature.

There are two types of neurons, visible and hiddeen

Page 23: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 24: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 25: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 26: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

This is applicable in Error Correction Learning.

Page 27: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

1)Pattern association problemPattern Recognition tasks by Feed Forward Networks

Page 28: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

1) Here every input ( training data ) is associated with an output.

2) So if an input ( test data ) , is close to any training data, like,

Then , Is associated with

3)But if the test data, is very far away from , training data, then Test data, will be associated with an output,

Note: Is very small

And not

4) System displays Accretive Behaviour.

5) Follows Feed Forward Network.

Page 29: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Ai=al + i1, i1 is small number.

Page 30: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

2)Pattern classification problem

Page 31: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

1)In Pattern Association problem, if a set of inputs map to an output, the size of output data set is smaller than input data set. Classes of inputs get a label.2)If a test data, which is close to any inputs ( training data ), in a class, it gets classified , to that class, for which there is a label.3) Here, test data is not associated , with output, but the class has a label, and test data is part of it.4) It creates Accretive behaviour.5) Follows Feed Forward Network.

Page 32: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 33: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

3)Pattern mapping

Page 34: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

1) Here output is a map of input.

2) So if any input ( test data ) , is close to any one training data, the output of test data, will be interpolation of output of training data, means they are in one range.3) Pattern Association and Pattern Classification are derived from Pattern Mapping. Show it by Interpolation.4) Pattern Mapping performs Generalization.5) Follows Feed Forward Network.

Page 35: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Auto Association Problem

Pattern Storage Problem

Pattern Environment Storage Problem

Pattern Recognition tasks by Feed Backward Networks

Page 36: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

1)Auto Association Networks

1) Inputs and Outputs are identical.2) Implementation has to be done by feed backward networks.3)Follows Feed Back Network.

Page 37: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

2)Pattern Storage problem.

The input and output are once again identical.

Three separate neurons are used to realize, the output.So output points and input points are different.Follows Feed Back Network.

Page 38: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

3)Pattern Environment Storage problem

If set of patterns, have certain probability, it is called as pattern environment Storage problem. Follows Feed Back Network.There is feedback, as to getOutput, we have to look at flip of states.

Page 39: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Pattern Recognition tasks by Competitive Learning.

If input patterns are replaced by new patterns, so that, the patterns get the output, over other patterns, it is called as temporary storageProblem. Follows CL. Some Input patterns want to reach output,

1)Temporary Storage problem

Page 40: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

2)Pattern Clustering problem

The test data is classified to the output, based on being near to first class. It creates Accretive Behaviour.Follows CL. Somehow test data, wants to enter testing data. A student, wants to enterer cluster of engg students.

Page 41: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Output is interpolative.Follows CL.Test Data, wants to somehow reach output.

Page 42: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

3)Feature Mapping problem

To cluster we need features by competitive Learning.Ex BP Algorithm.

Page 43: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Back Propagation Algorithm

Page 44: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 45: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 46: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 47: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 48: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Error at output neuron j.

Total Error Energy

1

2

Page 49: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Average of all energies ( at different discrete time intervals )

Activation Values.

3

4

Page 50: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Activation Function.

5

Page 51: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Consider

6

7

8

Page 52: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

9

10

Substituing 7 ,8 ,9 ,10 in 6 we get

11

Page 53: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Using Error Correction Rule

And LMS rule, we get

12

13

Using 11 in 13, we get

14

Page 54: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Where , the error gradient is given by,

15

The above , is to show another way of getting error gradient, to be used in 2nd part.

Page 55: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 56: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 57: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Using the expression in 15,

16

17

Total error at output layer18

Page 58: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

19

20

21

Page 59: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

22

23

24

Page 60: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

From 19

25

26

Substituting 20 in 16, we get

Page 61: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Using 14.

Page 62: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 63: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 64: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

BP Algorithm Summary

Page 65: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 66: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Virtues

Page 67: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 68: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.
Page 69: Supervised Learning. Teacher response: Emulation. Error: y1 – y2, where y1 is teacher response ( desired response, y2 is actual response ). Aim: To reduce.

Limitations ( where brain is better)