Top Banner
Hyperparameter Optimization 101 Alexandra Johnson Software Engineer, SigOpt
19

Hyperparameter Optimization 101

Jan 09, 2017

Download

Technology

SigOpt
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Hyperparameter Optimization 101

Hyperparameter Optimization 101Alexandra Johnson

Software Engineer, SigOpt

Page 2: Hyperparameter Optimization 101

What are Hyperparameters?

Page 3: Hyperparameter Optimization 101

Hyperparameters affect model performance

Page 4: Hyperparameter Optimization 101

How Do I Find The Best Hyperparameters?

Page 5: Hyperparameter Optimization 101

Step 1: Pick an Objective Metric

Classification models Accuracy

Regression models Root MSE

Page 6: Hyperparameter Optimization 101

Caveat: Cross Validate to Prevent Overfitting

Page 7: Hyperparameter Optimization 101

Cross Validation

4 5 60 1 2 3 7 8 9

4 5 6 7 8 90 1 2 3data

train validate metric

Page 8: Hyperparameter Optimization 101

Cross Validation

4 5 60 1 2 3 7 8 9

4 5 6 7 8 90 1 2 3data

train

6 7 91 2 4 5 0 3 8train

7 8 90 2 3 6 1 4 5train

metric

metric

metric

K ti

mes validate

validate

validate

Page 9: Hyperparameter Optimization 101

Grid Search Random Search Bayesian Optimization

Step 2: Pick an Optimization Strategy

Page 10: Hyperparameter Optimization 101

Step 3: Evaluate N Times

N Times

Page 11: Hyperparameter Optimization 101

What is the Best Hyperparameter Optimization Strategy?

Page 12: Hyperparameter Optimization 101

Primary Consideration: How Good are the “Best” Hyperparameters?

Page 13: Hyperparameter Optimization 101

“Best Found Value” Distributionsex

perim

ents

accuracy

Page 14: Hyperparameter Optimization 101

Secondary Consideration: How Much Time Do You Have?

Page 15: Hyperparameter Optimization 101

Number of Evaluations Required

Grid Search Random Search Bayesian Optimization

2 parameters 100 ?? 20-40

3 parameters 1,000 ?? 30-60

4 parameters 10,000 ?? 40-80

5 parameters 100,000 ?? 50-100

Page 16: Hyperparameter Optimization 101

SigOptEasy-to-use REST API, R, Java, Python Clients

Ensemble of Bayesian optimization techniques

Free trial, academic discount, we’re hiring!

Page 17: Hyperparameter Optimization 101

SigOpt Tutorial VideosVersus untuned models:

+315.2% accuracy with TensorFlow CNN

+49.2% accuracy with Xgboost + unsupervised features

Page 18: Hyperparameter Optimization 101

Learn MoreSee more at sigopt.com/research:● Blog posts● Papers● Videos

Page 19: Hyperparameter Optimization 101

Thank You!Twitter: @SigOpt

Email: [email protected]

Web: sigopt.com/getstarted