Solved – What does decay_steps mean in Tensorflow tf.train.exponential_decay

adamdeep learninggradient descentneural networkstensorflow

I am trying to implement an exponential learning rate decay with the Adam optimizer for a LSTM. I do not want the 'staircase = true' version. The decay_steps for me feels like the number of steps that the learning rate keeps constant. But I am not sure about this and Tensorflow has not stated it in their documentation. Any help is much appreciated.

Best Answer

As mentioned in the code of the function the relation of decay_steps with decayed_learning_rate is the following:

decayed_learning_rate = learning_rate *
                      decay_rate ^ (global_step / decay_steps)

Hence, you should set the decay_steps proportional to the global_step of the algorithm.