ICLR 2018
Skip to yearly menu bar Skip to main content


Workshop

Practical Hyperparameter Optimization

Stefan Falkner · ·

East Meeting Level 8 + 15 #2

Recently, the bandit-based strategy Hyperband (HB) was shown to yield good hyperparameter settings of deep neural networks faster than vanilla Bayesian optimization (BO). However, for larger budgets, HB is limited by its random search component, and BO works better. We propose to combine the benefits of both approaches to obtain a new practical state-of-the-art hyperparameter optimization method, which we show to consistently outperform both HB and BO on a range of problem types, including feed-forward neural networks, Bayesian neural networks, and deep reinforcement learning. Our method is robust and versatile, while at the same time being conceptually simple and easy to implement.

Live content is unavailable. Log in and register to view live content