The addition of an extra term, called a proportional factor reduces the convergence of the back propagation algorithm. So how to avoid local minima in Back propagation algorithm.
Asked
Active
Viewed 903 times
1 Answers
2
In local minimum a gradient of an error function is a zero vector - so backprop - which is using a gradient - cannot move your parameters any further and finishes training.

Marcin Możejko
- 39,542
- 10
- 109
- 120