I have a quick question regarding backpropagation. I am looking at the following:
http://www4.rgu.ac.uk/files/chapter3%20-%20bp.pdf
In this paper, it says to
The reason you need this is that you are calculating the derivative of the error function with respect to the neuron's inputs.
When you take the derivative via the chain rule, you need to multiply by the derivative of the neuron's activation function (which happens to be a sigmoid)
Here's the important math.
Calculate the derivative of the error on the neuron's inputs via the chain rule:
E = -(target - output)^2
dE/dinput = dE/doutput * doutput/dinput
Work out doutput/dinput:
output = sigmoid (input)
doutput/dinput = output * (1 - output) (derivative of sigmoid function)
therefore:
dE/dinput = 2 * (target - output) * output * (1 - output)