How to Calculate Confusion Matrix
What is Confusion Matrix?
Creates confusion matrix showing actual vs. predicted classifications. Basis for evaluation metrics.
Formula
Accuracy = (TP+TN) / total
- TP
- TN/(TN+FP) — TN/(TN+FP)
- TN
- TN value — Variable used in the calculation
Step-by-Step Guide
- 14 cells: TP (correct positive), FP (false positive), FN (false negative), TN (correct negative)
- 2Accuracy = (TP+TN) / total
- 3Sensitivity/Recall = TP/(TP+FN), Specificity = TN/(TN+FP)
- 4Precision = TP/(TP+FP)
Worked Examples
Input
TP/FP/TN/FN
Result
Metrics calc
Common Mistakes to Avoid
- ✕Using accuracy for imbalanced data (wrong)
- ✕Confusing sensitivity and specificity
- ✕Not balancing precision/recall tradeoff
Frequently Asked Questions
When use different metrics?
Accuracy: balanced classes; precision: minimize false positives; recall: minimize false negatives.
What about imbalanced classes?
Accuracy misleading; use precision, recall, F1-score, or AUC instead.
Ready to calculate? Try the free Confusion Matrix Calculator
Try it yourself →