Model-based Asynchronous Hyperparameter and Neural Architecture Search
Mar 1, 2020·
,,·
0 min read
Aaron Klein
Equal contribution
Louis Tiao
Equal contribution
,Thibaut Lienart
Cédric Archambeau
Matthias Seeger

Abstract
We introduce a model-based asynchronous multi-fidelity method for hyperparameter and neural architecture search that combines the strengths of asynchronous Hyperband and Gaussian process-based Bayesian optimization. At the heart of our method is a probabilistic model that can simultaneously reason across hyperparameters and resource levels, and supports decision-making in the presence of pending evaluations. We demonstrate the effectiveness of our method on a wide range of challenging benchmarks, for tabular data, image classification and language modelling, and report substantial speed-ups over current state-of-the-art methods. Our new methods, along with asynchronous baselines, are implemented in a distributed framework that is open-sourced alongside this publication.
Type
Hyperparameter Optimization
Neural Architecture Search
Multi-Fidelity Optimization
AutoML
Bayesian Optimization
Gaussian Processes
Parallel Computing
Machine Learning
Authors

Authors
Research Scientist
My name is Louis Tiao, and I graduated from one of Australia’s
top engineering schools with really good grades.
Now, I’m using my knowledge to help up-and-coming tech companies
make it in this competitive world.