Search Methods

Determined supports a variety of hyperparameter search algorithms. Aside from the single searcher, a searcher runs multiple trials and decides the hyperparameter values to use in each trial. Every searcher is configured with the name of the validation metric to optimize (via the metric field), in addition to other searcher-specific options. For example, the (state-of-the-art) adaptive_asha searcher, suitable for larger experiments with many trials, is configured with the maximum number of trials to run, the maximum training length allowed per trial, and the maximum number of trials that can be worked on simultaneously:

  name: "adaptive_asha"
  metric: "validation_loss"
  max_trials: 16
      epochs: 1
  max_concurrent_trials: 8

For details on the supported searchers and their respective configuration options, refer to Hyperparameter Tuning.

That’s it! After submitting an experiment, users can easily see the best validation metric observed across all trials over time in the WebUI. After the experiment has completed, they can view the hyperparameter values for the best-performing trials and then export the associated model checkpoints for downstream serving.


Other Supported Search Methods

Determined also supports other common hyperparameter search algorithms:

  • Single is appropriate for manual hyperparameter tuning, as it trains a single hyperparameter configuration.

  • Grid brute force evaluates all possible hyperparameter configurations and returns the best.

  • Random evaluates a set of hyperparameter configurations chosen at random and returns the best.

  • Population-based training (PBT) begins as random search but periodically replaces low-performing hyperparameter configurations with ones near the high-performing points in the hyperparameter space.