Efficient Wasserstein Natural Gradients for Reinforcement Learning

Ted Moskovitz · Michael Arbel · Ferenc Huszar · Arthur Gretton


Keywords: [ reinforcement learning ] [ optimization ]

[ Abstract ]
[ Slides
[ Paper ]
Wed 5 May 5 p.m. PDT — 7 p.m. PDT


A novel optimization approach is proposed for application to policy gradient methods and evolution strategies for reinforcement learning (RL). The procedure uses a computationally efficient \emph{Wasserstein natural gradient} (WNG) descent that takes advantage of the geometry induced by a Wasserstein penalty to speed optimization. This method follows the recent theme in RL of including divergence penalties in the objective to establish trust regions. Experiments on challenging tasks demonstrate improvements in both computational cost and performance over advanced baselines.

Chat is not available.