regressionPairs takes the training and test sets in data and
processes it using pre (if supplied). regressor is called to
induce a regression model for each pair of algorithms to predict the performance
difference between them. If combine is not supplied, the best overall
algorithm is determined by summing the performance differences over all pairs
for each algorithm and ranking them by this sum. The algorithm with the largest
value is chosen. If it is supplied, it is assumed to be an mlr classifier. This classifier is
passed the original features and the predictions for each pair of algorithms. If
the classifier supports weights, the performance difference between the best and
the worst algorithm is passed as weight.The aggregated score for each algorithm quantifies how much better it is than
the other algorithms, where bigger values are better. Positive numbers denote
that the respective algorithm usually exhibits better performance than most of
the other algorithms, while negative numbers denote that it is usually worse.
The evaluation across the training and test sets will be parallelized
automatically if a suitable backend for parallel computation is loaded.
Training this model can take a very long time. Given n algorithms,
choose(n, 2) * n models are trained and evaluated. This is significantly
slower than the other approaches that train a single model or one for each
algorithm.