Multilayer perceptron
General
A multilayer perceptron is a feedforward artificial neural network. This means the signal inside the neural network flows from input layer passing hidden layers to output layer. While training the error correction of neural weights are done in the opposite direction. This is done by the backpropagation algorithm.
Activation
At first a cumulative input is calculated by the following equation:
Sigmoid activation function
Tangens hyperbolicus activation function
using output range between -1 and 1, or
using output range between 0 and 1.
- cumulative input
- weight of input
- value of input
- number of inputs
- number of neuron
Error of neural network
If the neural network is initialized by random weights it has of course not the expected output. Therefore training is necessary. While supervised training known inputs and their corresponded output values are presented to the network. So it is possible to compare the real output with the desired output. The error is described as the following algorithm:
- network error
- count of input patterns
- desired output
- calculated output
Backpropagation
The learning algorithm of a single layer perceptron is easy compared to a multilayer perceptron. The reason is that just the output layer is directly connected to the output, but not the hidden layers. Therefore the calculation of the right weights of the hidden layers is difficult mathematically. To get the right delta value for changing the weights of hidden neuron is described in the following equation:
- network error
- delta value of neuron connection to
- learning rate
- the error of neuron
- input of neuron
- desired output of output neuron
- real output of output neuron .
Programming solution of backpropagation
In this PHP implementation of multilayer perceptron the following algorithm is used for weight changes in hidden layers:
- learning rate
- momentum
- neuron k
- neuron l
- weight m
- input
- output
- count of neurons
To avoid overfitting of neural network the training procedure is finished if real output value has a fault tolerance of 1 per cent of desired output value.
Choosing learning rate and momentum
The proper choosing of learning rate () and momentum () is done by experience. Both values have a range between 0 and 1. This PHP implementation uses a default value of 0.5 for and 0.95 for . Theses factors can be change by runtime.