Revision 24f5148a7ae3e175a873f0210e56e2b70861a155 authored by dependabot[bot] on 23 March 2020, 16:12:45 UTC, committed by GitHub on 23 March 2020, 16:12:45 UTC
Bumps [tensorflow](https://github.com/tensorflow/tensorflow) from 1.15.0 to 1.15.2.
- [Release notes](https://github.com/tensorflow/tensorflow/releases)
- [Changelog](https://github.com/tensorflow/tensorflow/blob/master/RELEASE.md)
- [Commits](https://github.com/tensorflow/tensorflow/compare/v1.15.0...v1.15.2)

Signed-off-by: dependabot[bot] <support@github.com>

Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
1 parent 932648f
Raw File
algorithm-settings.md
# Hyperparameter Tuning Algorithms

Table of Contents
=================

   * [Hyperparameter Tuning Algorithms](#hyperparameter-tuning-algorithms)
   * [Table of Contents](#table-of-contents)
      * [Grid Search](#grid-search)
         * [<a href="https://chocolate.readthedocs.io" rel="nofollow">Chocolate</a>](#chocolate)
      * [Random Search](#random-search)
         * [<a href="http://hyperopt.github.io/hyperopt/" rel="nofollow">Hyperopt</a>](#hyperopt)
      * [TPE](#tpe)
         * [<a href="http://hyperopt.github.io/hyperopt/" rel="nofollow">Hyperopt</a>](#hyperopt-1)
      * [Bayesian Optimization](#bayesian-optimization)
         * [<a href="https://github.com/scikit-optimize/scikit-optimize">scikit-optimize</a>](#scikit-optimize)
      * [References](#references)

Created by [gh-md-toc](https://github.com/ekalinin/github-markdown-toc)

<!-- ## Quasi Random Search

QuasiRandom sampling ensures a much more uniform exploration of the search space than traditional pseudo random. Thus, quasi random sampling is preferable when not all variables are discrete, the number of dimensions is high and the time required to evaluate a solution is high.

### [Chocolate](https://chocolate.readthedocs.io)

Algorithm name in katib is `chocolate-quasirandom`. -->

<!-- ## CMAES

CMAES search is one of the most powerful black-box optimization algorithm. However, it requires a significant number of model evaluation (in the order of 10 to 50 times the number of dimensions) to converge to an optimal solution. This search method is more suitable when the time required for a model evaluation is relatively low.

###  [Chocolate](https://chocolate.readthedocs.io)

Algorithm name in katib is `chocolate-CMAES`. -->

For information about the hyperparameter tuning algorithms and neural
architecture search implemented or integrated in Katib, see the detailed guide
to [configuring and running a Katib 
experiment](https://kubeflow.org/docs/components/hyperparameter-tuning/experiment/).
back to top