Momentum Backpropagation
Momentum Backpropagation allows a dynamic change of the learning rate in a Neural Network by including a memory term in the weight update. This will lead to larger steps in the direction of similar Gradients and smaller steps in oscillating or zigzagging gradients. The discounting factor should be between zero and one.
Via the Geometric Series we get for similar Gradients a way to speed up the steps.
And for oscillating gradients () we get a way to slow down the steps.
This method is especially helpful when we have flat regions in the parameters of the network.
It is used in the Adam optimizer.