Skip to yearly menu bar Skip to main content


Poster

Value Propagation Networks

Nantas Nardelli · Gabriel Synnaeve · Zeming Lin · Pushmeet Kohli · Philip Torr · Nicolas Usunier

Great Hall BC #7

Keywords: [ learning to plan ] [ value iteration ] [ convolutional neural networks ] [ reinforcement learning ] [ navigation ]


Abstract:

We present Value Propagation (VProp), a set of parameter-efficient differentiable planning modules built on Value Iteration which can successfully be trained using reinforcement learning to solve unseen tasks, has the capability to generalize to larger map sizes, and can learn to navigate in dynamic environments. We show that the modules enable learning to plan when the environment also includes stochastic elements, providing a cost-efficient learning system to build low-level size-invariant planners for a variety of interactive navigation problems. We evaluate on static and dynamic configurations of MazeBase grid-worlds, with randomly generated environments of several different sizes, and on a StarCraft navigation scenario, with more complex dynamics, and pixels as input.

Live content is unavailable. Log in and register to view live content