Get hyperparameter values
get_hyperparameter_defaults(
  models = get_supported_models(),
  n = 100,
  k = 10,
  model_class = "classification"
)get_random_hyperparameters(
  models = get_supported_models(),
  n = 100,
  k = 10,
  tune_depth = 5,
  model_class = "classification"
)
Named list of data frames. Each data frame corresponds to an algorithm, and each column in each data fram corresponds to a hyperparameter for that algorithm. This is the same format that should be provided to
tune_models(hyperparameters = ) to specify hyperparameter values.
which algorithms?
Number observations
Number features
"classification" or "regression"
How many combinations of hyperparameter values?
Get hyperparameters for model training.
 get_hyperparameter_defaults returns a list of 1-row data frames
 (except for glm, which is a 10-row data frame) with default hyperparameter
 values that are used by flash_models.
 get_random_hyperparameters returns a list of data frames with
 combinations of random values of hyperparameters to tune over in
 tune_models; the number of rows in the data frames is given by
 `tune_depth`.
For get_hyperparameter_defaults
 XGBoost defaults are from caret and XGBoost documentation:
 eta = 0.3, gamma = 0, max_depth = 6, subsample = 0.7,
 colsample_bytree = 0.8, min_child_weight = 1, and nrounds = 50.
 Random forest defaults are from Intro to
 Statistical Learning and caret: mtry = sqrt(k), splitrule = "extratrees",
 min.node.size = 1 for classification, 5 for regression.
 glm defaults are
 from caret: alpha = 1, and because glmnet fits sequences of lambda nearly as
 fast as an individual value, lambda is a sequence from 1e-4 to 8.
models for model and hyperparameter details