References – Sensitivity vs. Specificity vs. Recall: Comprehensive Comparison and Analysis

confusion matrixprecision-recallreferencessensitivity-specificity

Given a binary confusion matrix with true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN), what are the formulas for sensitivity, specificity, and recall?

I'm coming across many conflicting answers online. For instance, wikipedia says:

enter image description here

But in a textbook — "Data Science for Business: What You Need to Know about Data Mining and Data-Analytic Thinking" (Provost & Fawcett, 2013) — it says:

enter image description here

Which one is correct? Is there an authoritative reference on this?

Best Answer

The sensitivity of the model is the rate at the class of interest is predicted correctly for all samples having the class. Wikipedia is correct. To quote the Elements of Statistical Learning by Friedman et al. "Sensitivity: probability of predicting disease given true state is disease." the book has a nice worked example in Ch. 9. Similarlly in Probabilistic Machine Learning: An Introduction by Kevin Patrick Murphy: "we can compute the true positive rate (TPR), also known as the sensitivity, recall or hit rate, by using (...)". Both books are well-accepted as authoritative references in ML. The "Data Science for Business" book had some copywriting and/or proofreading error on this.

Related Question