## S3 method for class 'formula':
blackboost(formula, data = list(), weights = NULL, ...)
## S3 method for class 'matrix':
blackboost(x, y, weights = NULL, ...)
blackboost_fit(object, tree_controls =
ctree_control(teststat = "max",
testtype = "Teststatistic",
mincriterion = 0,
maxdepth = 2),
fitmem = ctree_memory(object, TRUE), family = GaussReg(),
control = boost_control(), weights = NULL)
boost_data
, see boost_dpp
."TreeControl "
, which can be
obtained using ctree_control
.
Defines hyper-parameters for the trees whiTreeFitMemory
."boost_family "
,
implementing the negative gradient corresponding
to the loss function to be optimized. By default,
squared error loss
boost_control
which defines the hyper-parameters of the
boosting algorithm.gbm
. The
main difference is that arbitrary loss functions to be optimized
can be specified via the family
argument to blackboost
whereas
gbm
uses hard-coded loss functions.
Moreover, the base learners (conditional
inference trees, see ctree
) are a little bit more flexible.The regression fit is a black box prediction machine and thus hardly interpretable.
Usually, the formula based interface blackboost
should be used.
When necessary (for example for cross-validation), function
blackboost_fit
operating on objects of class boost_data
is faster alternative.
Greg Ridgeway (1999), The state of boosting. Computing Science and Statistics, 31, 172--181.
Peter Buhlmann and Torsten Hothorn (2007), Boosting algorithms: regularization, prediction and model fitting. Statistical Science, 22(4), 477--505.
Torsten Hothorn, Kurt Hornik and Achim Zeileis (2006). Unbiased Recursive Partitioning: A Conditional Inference Framework. Journal of Computational and Graphical Statistics, 15(3), 651--674.
glmboost
for boosted linear models and
gamboost
for boosted smooth models. See
cvrisk
for cross-validated stoping iteration.
Furthermore see boost_control
, Family
and
methods
### a simple two-dimensional example: cars data
cars.gb <- blackboost(dist ~ speed, data = cars,
control = boost_control(mstop = 50))
cars.gb
### plot fit
plot(dist ~ speed, data = cars)
lines(cars$speed, predict(cars.gb), col = "red")
Run the code above in your browser using DataLab