Solved – RMSprop difference rho and decay in Tensorflow

deep learningtensorflow

As this post showed nicely, there is a difference between rho and the decay in RMSprop. I can't clearly see what tensorflows RMSprop parameter decays stands for. Is this the learning rate decay? And if so, where is the rho?

Best Answer

In the post that you added they talk about RMSprop implementation from keras (it's included in tensorflow), but the other link that you added is original implementation for the RMSProp from tensorflow. Tensorflow's implementation has parameter called decay, but it's the same as rho in the keras library. But in the RMSProp implantation from keras the decay parameter reduced learning rate overtime.