Create control object for hyperparameter tuning with MBO.
makeTuneControlMBO(same.resampling.instance = TRUE, impute.val = NULL, learner = NULL, mbo.control = NULL, tune.threshold = FALSE, tune.threshold.args = list(), continue = FALSE, log.fun = "default", final.dw.perc = NULL, budget = NULL, mbo.design = NULL)
logical(1)) Should the same resampling instance be used for all evaluations to reduce variance? Default is
(numeric) If something goes wrong during optimization (e.g. the learner crashes), this value is fed back to the tuner, so the tuning algorithm does not abort. It is not stored in the optimization path, an NA and a corresponding error message are logged instead. Note that this value is later multiplied by -1 for maximization measures internally, so you need to enter a larger positive value for maximization here as well. Default is the worst obtainable value of the performance measure you optimize for when you aggregate by mean value, or
Infinstead. For multi-criteria optimization pass a vector of imputation values, one for each of your measures, in the same order as your measures.
NULL) The surrogate learner: A regression learner to model performance landscape. For the default,
NULL, mlrMBO will automatically create a suitable learner based on the rules described in mlrMBO::makeMBOLearner.
logical(1)) Should the threshold be tuned for the measure at hand, after each hyperparameter evaluation, via tuneThreshold? Only works for classification if the predict type is “prob”. Default is
logical(1)) Resume calculation from previous run using mlrMBO::mboContinue? Requires “save.file.path” to be set. Note that the ParamHelpers::OptPath in the mlrMBO::OptResult will only include the evaluations after the continuation. The complete OptPath will be found in the slot
character(1)) Function used for logging. If set to “default” (the default), the evaluated design points, the resulting performances, and the runtime will be reported. If set to “memory” the memory usage for each evaluation will also be displayed, with
character(1)small increase in run time. Otherwise
character(1)function with arguments
prev.stageis expected. The default displays the performance measures, the time needed for evaluating, the currently used memory and the max memory ever used before (the latter two both taken from gc). See the implementation for details.
boolean) If a Learner wrapped by a makeDownsampleWrapper is used, you can define the value of
dw.percwhich is used to train the Learner with the final parameter setting found by the tuning. Default is
NULLwhich will not change anything.
integer(1)) Maximum budget for tuning. This value restricts the number of function evaluations.
NULL) Initial design as data frame. If the parameters have corresponding trafo functions, the design must not be transformed before it is passed! For the default,
NULL, a default design is created like described in mlrMBO::mbo.
Bernd Bischl, Jakob Richter, Jakob Bossek, Daniel Horn, Janek Thomas and Michel Lang; mlrMBO: A Modular Framework for Model-Based Optimization of Expensive Black-Box Functions, Preprint: https://arxiv.org/abs/1703.03373 (2017).