Optimizing performance of Deep Neural Networks by tuning its Hyper-parameters

Optimizing performance of Deep Neural Networks by tuning its Hyper-parameters

The performance of a Deep Neural Network depends on a wide range of hyperparameters which makes manual tuning an extremely challenging and tedious task. The search space is vast and it is not feasible to try all combinations, also a hit and trial method does not scale. Hence a hyperparameter tuning strategy using efficient algorithms needs to be adopted to find the optimal hyperparameters for a Deep Neural Network.

The approach for Hyperparameter Optimization entails :

  • Define the search space for the hyperparameters for sampling
  • Select an algorithm which the tuner would use.
  • Generate a new set of hyperparameter values from the search space for each trial using the tuner
  • Build and evaluate the models using the hyperparameters and record the metrics
  • Finally, a good set of hyperparameter values is found which results in good performance

In this blog a comparative study of several popular hyperparameter optimization algorithms like Random Search,Bayesian Optimization and Hyperband has been performed. These algorithms can be experimented with to improve the model’s efficiency.

Adding a data source

comments powered by Disqus