powered by
Measure to compare true observed labels with predicted labels in binary classification tasks.
fp(truth, response, positive, ...)
Performance value as numeric(1).
numeric(1)
(factor()) True (observed) labels. Must have the exactly same two levels and the same length as response.
factor()
response
(factor()) Predicted response labels. Must have the exactly same two levels and the same length as truth.
truth
(character(1)) Name of the positive class.
character(1))
(any) Additional arguments. Currently ignored.
any
Type: "binary"
"binary"
Range: \([0, \infty)\)
Minimize: TRUE
TRUE
Required prediction: response
This measure counts the false positives (type 1 error), i.e. the number of predictions indicating a positive class label while in fact it is negative.
https://en.wikipedia.org/wiki/Template:DiagnosticTesting_Diagram
Other Binary Classification Measures: auc(), bbrier(), dor(), fbeta(), fdr(), fnr(), fn(), fomr(), fpr(), mcc(), npv(), ppv(), prauc(), tnr(), tn(), tpr(), tp()
auc()
bbrier()
dor()
fbeta()
fdr()
fnr()
fn()
fomr()
fpr()
mcc()
npv()
ppv()
prauc()
tnr()
tn()
tpr()
tp()
set.seed(1) lvls = c("a", "b") truth = factor(sample(lvls, 10, replace = TRUE), levels = lvls) response = factor(sample(lvls, 10, replace = TRUE), levels = lvls) fp(truth, response, positive = "a")
Run the code above in your browser using DataLab