Tune

How to tune deep learning models?

Summary

  • Choosing which hyper-parameters to optimize is not an easy task since some are more sensitive than others and are dependent upon the choice of model.

    • Low sensitivity: Optimizer, batch size, non-linearity.

    • Medium sensitivity: weight initialization, model depth, layer parameters, weight of regularization.

    • High sensitivity: learning rate, annealing schedule, loss function, layer size.

  • Method 1 is manual optimization:

    • For a skilled practitioner, this may require the least amount of computation to get good results.

    • However, the method is time-consuming and requires a detailed understanding of the algorithm.

  • Method 2 is grid search:

    • Grid search is super simple to implement and can produce good results.

    • Unfortunately, it’s not very efficient since we need to train the model on all cross-combinations of the hyper-parameters. It also requires prior knowledge about the parameters to get good results.

  • Method 3 is random search:

    • Random search is also easy to implement and often produces better results than grid search.

    • But it is not very interpretable and may also require prior knowledge about the parameters to get good results.

  • Method 4 is coarse-to-fine search:

    • This strategy helps you narrow in only on very high performing hyper-parameters and is a common practice in the industry.

    • The only drawback is that it is somewhat a manual process.

  • Method 5 is Bayesian optimization search:

    • Bayesian optimization is generally the most efficient hands-off way to choose hyper-parameters.

    • But it’s difficult to implement from scratch and can be hard to integrate with off-the-shelf tools.

Last updated