Neural networks, connectionism, parallel distributed processing
Based on abstract view of the neuron
The connections determine the function of the network
Connections can be formed by learning and do not need to be programmed
Logic gates
Computers have electronic elements that implement ‘logic gates; and with these you can build and run programs
McCullock-Pitts neuron - assumptions
McCullock-Pitts neuron - 1 major thing the didn’t take into account
Real neurons are very noisy
So computation in the brain is fault tolerant, which doesn’t work in a computational model (that would give an error)
i.e., the brain does not work like a turing machine
so neural networks abstract strongly from the details of real neurons
neural networks abstract strongly from the details of real neurons
in what way do they differ
Error-correcting learning
form of supervised learning
Perceptron
Original perceptron had only 2 layers (input and output layer)
Limitations of the perceptron
Perceptron convergence theorem
If a pattern set can be represented by a two-layer perceptron, the perceptron learning rule will always be able to find some correct weights
o So if it can, it will.
o Does not say anything about how fast. Could be a slow process. But it will find it in the end.
needed for error backpropagation
Characteristics of backpropagation
backprop trick
We have a node h in the hidden layer
We go to the error signal on the output layer that is calculated for each node
o Error = the difference between the target and the spontaneous output
We take all those errors in the output layer and add them up. This is the error we have for our hidden layer
o Can be positive and negative
Not biologically plausible because axons only work in 1 direction.
Backpropagation algorithm in rules
Disadvantages backprop
advantages backprop
gradient descent in error space (= find the steepest slope down), but
i. does not guarantee high performance
ii. does not prevent getting stuck in local minima (unlike perceptron that will find the solution if there is one)
iii. the learning rule is complicated and tends to slow down