Hyperband: a novel bandit-based approach to hyperparameter optimization

From MaRDI portal
Publication:72746

DOI10.48550/ARXIV.1603.06560zbMATH Open1468.68204arXiv1603.06560MaRDI QIDQ72746FDOQ72746


Authors: Lisha Li, Kevin Jamieson, Giulia DeSalvo, Afshin Rostamizadeh, Ameet Talwalkar, Lisha Li, Kevin Jamieson, Giulia Desalvo, Afshin Rostamizadeh, Ameet Talwalkar Edit this on Wikidata


Publication date: 21 March 2016

Abstract: Performance of machine learning algorithms depends critically on identifying a good set of hyperparameters. While recent approaches use Bayesian optimization to adaptively select configurations, we focus on speeding up random search through adaptive resource allocation and early-stopping. We formulate hyperparameter optimization as a pure-exploration non-stochastic infinite-armed bandit problem where a predefined resource like iterations, data samples, or features is allocated to randomly sampled configurations. We introduce a novel algorithm, Hyperband, for this framework and analyze its theoretical properties, providing several desirable guarantees. Furthermore, we compare Hyperband with popular Bayesian optimization methods on a suite of hyperparameter optimization problems. We observe that Hyperband can provide over an order-of-magnitude speedup over our competitor set on a variety of deep-learning and kernel-based learning problems.


Full work available at URL: https://arxiv.org/abs/1603.06560




Recommendations




Cites Work


Cited In (49)

Uses Software





This page was built for publication: Hyperband: a novel bandit-based approach to hyperparameter optimization

Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q72746)