In-Person Poster presentation / top 5% paper
Near-optimal Policy Identification in Active Reinforcement Learning
Xiang Li · Viraj Mehta · Johannes Kirschner · Ian Char · Willie Neiswanger · Jeff Schneider · Andreas Krause · Ilija Bogunovic
MH1-2-3-4 #101
Keywords: [ kernelized least-squares value iteration ] [ contextual bayesian optimization ] [ reinforcement learning ] [ Reinforcement Learning ]
Many real-world reinforcement learning tasks require control of complex dynamical systems that involve both costly data acquisition processes and large state spaces. In cases where the expensive transition dynamics can be readily evaluated at specified states (e.g., via a simulator), agents can operate in what is often referred to as planning with a \emph{generative model}. We propose the AE-LSVI algorithm for best policy identification, a novel variant of the kernelized least-squares value iteration (LSVI) algorithm that combines optimism with pessimism for active exploration (AE). AE-LSVI provably identifies a near-optimal policy \emph{uniformly} over an entire state space and achieves polynomial sample complexity guarantees that are independent of the number of states. When specialized to the recently introduced offline contextual Bayesian optimization setting, our algorithm achieves improved sample complexity bounds. Experimentally, we demonstrate that AE-LSVI outperforms other RL algorithms in a variety of environments when robustness to the initial state is required.