Get hyperparameter values
get_hyperparameter_defaults(models = get_supported_models(), n = 100,
  k = 10, model_class = "classification")get_random_hyperparameters(models = get_supported_models(), n = 100,
  k = 10, tune_depth = 5, model_class = "classification")
which algorithms?
Number observations
Number features
"classification" or "regression"
How many combinations of hyperparameter values?
Named list of data frames. Each data frame corresponds to an
 algorithm, and each column in each data fram corresponds to a hyperparameter
 for that algorithm. This is the same format that should be provided to
 tune_models(hyperparameters = ) to specify hyperparameter values.
Get hyperparameters for model training.
 get_hyperparameter_defaults returns a list of 1-row data frames
 (except for glm, which is a 10-row data frame) with default hyperparameter
 values that are used by flash_models.
 get_random_hyperparameters returns a list of data frames with
 combinations of random values of hyperparameters to tune over in
 tune_models; the number of rows in the data frames is given by
 `tune_depth`.
For get_hyperparameter_defaults
 XGBoost defaults are from caret and XGBoost documentation:
 eta = 0.3, gamma = 0, max_depth = 6, subsample = 0.7,
 colsample_bytree = 0.8, min_child_weight = 1, and nrounds = 50.
 Random forest defaults are from Intro to
 Statistical Learning and caret: mtry = sqrt(k), splitrule = "extratrees",
 min.node.size = 1 for classification, 5 for regression.
 glm defaults are
 from caret: alpha = 1, and because glmnet fits sequences of lambda nearly as
 fast as an individual value, lambda is a sequence from 1e-4 to 8.
models for model and hyperparameter details