You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
weights[n - 1] = weights[n - 1].plus(deltas);
// Calculate and set target for previous (next) layer
SimpleMatrix previousError = weights[n - 1].transpose().mult(errors);
target = previousError.plus(layers[n - 1]);
you calculate the target for the next (lower) level using the matrix previousError.
That uses the weights weights[n - 1] which had previously been altered with the correction factor.
What you require here is the unaltered weight to make the lower layer of the network backpropagation dependent on the original error with the original weight, not the altered weight.
From my point of view this can be fixed by moving the previousError line above the weights computation, like:
in your lines https://github.com/kim-marcel/basic_neural_network/blob/master/src/main/java/basicneuralnetwork/NeuralNetwork.java#L165-L169
you calculate the target for the next (lower) level using the matrix
previousError
.That uses the weights
weights[n - 1]
which had previously been altered with the correction factor.What you require here is the unaltered weight to make the lower layer of the network backpropagation dependent on the original error with the original weight, not the altered weight.
From my point of view this can be fixed by moving the
previousError
line above the weights computation, like:The text was updated successfully, but these errors were encountered: