powered by
Decision Forest algorithm: Model training
DF_train(X, Y, stop_step = 5, Max_tree = 20, min_split = 10, cp = 0.1, Filter = F, p_val = 0.05, Method = "bACC", Quiet = T, Grace_val = 0.05, imp_accu_val = 0.01, imp_accu_criteria = F)
Training Dataset
Training data endpoint
How many extra step would be processed when performance not improved, 1 means one extra step
Maximum tree number in Forest
minimum leaves in tree nodes
parameters to pruning decision tree, default is 0.1
doing feature selection before training
P-value threshold measured by t-test used in feature selection, default is 0.05
Which is used for evaluating training process. MIS: Misclassification rate; ACC: accuracy
if TRUE (default), don't show any message during the process
Grace Value in evaluation: the next model should have a performance (Accuracy, bACC, MCC) not bad than previous model with threshold
improvement in evaluation: adding new tree should improve the overall model performance (Accuracy, bACC, MCC) by threshold
if TRUE, model must have improvement in accumulated accuracy
.$accuracy: Overall training accuracy
.$pred: Detailed training prediction (fitting)
.$detail: Detailed usage of Decision tree Features/Models and their performances
.$models: Constructed (list of) Decision tree models
.$Method: pass evaluating Methods used in training
.$cp: pass cp value used in training decision trees
# NOT RUN { ##data(iris) X = iris[,1:4] Y = iris[,5] names(Y)=rownames(X) used_model = DF_train(X,factor(Y)) # }
Run the code above in your browser using DataLab