RMSprop modifies the AdaGrad algorithm by taking the EWMA of previous squared gradients. It has the moving average parameter, ρ, which controls the length and scale of the moving average. This is one of the most successful algorithms for deep neural network training.