Skip to yearly menu bar Skip to main content


Poster

Value Propagation Networks

Nantas Nardelli · Gabriel Synnaeve · Zeming Lin · Pushmeet Kohli · Philip Torr · Nicolas Usunier

Keywords: [ navigation ] [ reinforcement learning ] [ convolutional neural networks ] [ value iteration ] [ learning to plan ]

[ ]
[ PDF
2019 Poster

Abstract:

We present Value Propagation (VProp), a set of parameter-efficient differentiable planning modules built on Value Iteration which can successfully be trained using reinforcement learning to solve unseen tasks, has the capability to generalize to larger map sizes, and can learn to navigate in dynamic environments. We show that the modules enable learning to plan when the environment also includes stochastic elements, providing a cost-efficient learning system to build low-level size-invariant planners for a variety of interactive navigation problems. We evaluate on static and dynamic configurations of MazeBase grid-worlds, with randomly generated environments of several different sizes, and on a StarCraft navigation scenario, with more complex dynamics, and pixels as input.

Chat is not available.