Resilient Backpropagation (Rprop) for Batch-learning in TensorFlow
Ciprian Florescu · Christian Igel
Abstract
The resilient backpropagation (Rprop) algorithms are fast and accurate batch learning methods for neural networks. We describe their implementation in the popular machine learning framework TensorFlow. We present the first empirical evaluation of Rprop for training recurrent neural networks with gated recurrent units. In our experiments, Rprop with default hyperparameters outperformed vanilla steepest descent as well as the optimization algorithms RMSprop and Adam even if their hyperparameters were tuned.
Video
Chat is not available.
Successful Page Load