## Practical Real Time Recurrent Learning with a Sparse Approximation

### Jacob Menick · Erich Elsen · Utku Evci · Simon Osindero · Karen Simonyan · Alex Graves

Keywords: [ bptt ] [ rtrl ] [ real time recurrent learning ] [ forward mode ] [ biologically plausible ] [ backpropagation ] [ recurrent neural networks ]

[ Abstract ]
[ Paper ]
Tue 4 May 1 a.m. PDT — 3 a.m. PDT

Spotlight presentation: Oral Session 10
Thu 6 May 3 a.m. PDT — 5:40 a.m. PDT

Abstract: Recurrent neural networks are usually trained with backpropagation through time, which requires storing a complete history of network states, and prohibits updating the weights "online" (after every timestep). Real Time Recurrent Learning (RTRL) eliminates the need for history storage and allows for online weight updates, but does so at the expense of computational costs that are quartic in the state size. This renders RTRL training intractable for all but the smallest networks, even ones that are made highly sparse. We introduce the Sparse n-step Approximation (SnAp) to the RTRL influence matrix. SnAp only tracks the influence of a parameter on hidden units that are reached by the computation graph within $n$ timesteps of the recurrent core. SnAp with $n=1$ is no more expensive than backpropagation but allows training on arbitrarily long sequences. We find that it substantially outperforms other RTRL approximations with comparable costs such as Unbiased Online Recurrent Optimization. For highly sparse networks, SnAp with $n=2$ remains tractable and can outperform backpropagation through time in terms of learning speed when updates are done online.

Chat is not available.