Matthews correlation coefficient
The Matthews correlation coefficient is used in machine learning as a measure of the quality of binary (two-class) classifications, introduced by biochemist Brian W. Matthews in 1975.[1] It takes into account true and false positives and negatives and is generally regarded as a balanced measure which can be used even if the classes are of very different sizes. The MCC is in essence a correlation coefficient between the observed and predicted binary classifications; it returns a value between −1 and +1. A coefficient of +1 represents a perfect prediction, 0 no better than random prediction and −1 indicates total disagreement between prediction and observation. The statistic is also known as the phi coefficient. MCC is related to the chi-square statistic for a 2×2 contingency table
where n is the total number of observations.
While there is no perfect way of describing the confusion matrix of true and false positives and negatives by a single number, the Matthews correlation coefficient is generally regarded as being one of the best such measures.[2] Other measures, such as the proportion of correct predictions (also termed accuracy), are not useful when the two classes are of very different sizes. For example, assigning every object to the larger set achieves a high proportion of correct predictions, but is not generally a useful classification.
The MCC can be calculated directly from the confusion matrix using the formula:
In this equation, TP is the number of true positives, TN the number of true negatives, FP the number of false positives and FN the number of false negatives. If any of the four sums in the denominator is zero, the denominator can be arbitrarily set to one; this results in a Matthews correlation coefficient of zero, which can be shown to be the correct limiting value.
The original formula as given by Matthews was:[1]
This is equal to the formula given above. As a correlation coefficient, the Matthews correlation coefficient is the geometric mean of the regression coefficients of the problem and its dual. The component regression coefficients of the Matthews correlation coefficient are Markedness (Δp) and Youden's J statistic (Informedness or deltap').[2][3] Markedness and Informedness correspond to different directions of information flow and generalize Youden's J statistic, the deltap statistics and (as their geometric mean) the Matthews Correlation Coefficient to more than two classes.[2]
Confusion Matrix
<templatestyles src="Module:Hatnote/styles.css"></templatestyles>
|
Let us define an experiment from P positive instances and N negative instances for some condition. The four outcomes can be formulated in a 2×2 contingency table or confusion matrix, as follows:
Predicted condition | |||||
Total population | Predicted Condition positive | Predicted Condition negative | Prevalence = <templatestyles src="Sfrac/styles.css" />Σ Condition positive/Σ Total population | ||
True condition |
condition positive |
True positive | False Negative (Type II error) |
True positive rate (TPR), Sensitivity, Recall = <templatestyles src="Sfrac/styles.css" />Σ True positive/Σ Condition positive | False negative rate (FNR), Miss rate = <templatestyles src="Sfrac/styles.css" />Σ False negative/Σ Condition positive |
condition negative |
False Positive (Type I error) |
True negative | False positive rate (FPR), Fall-out = <templatestyles src="Sfrac/styles.css" />Σ False positive/Σ Condition negative | True negative rate (TNR), Specificity (SPC) = <templatestyles src="Sfrac/styles.css" />Σ True negative/Σ Condition negative | |
Accuracy (ACC) = <templatestyles src="Sfrac/styles.css" />Σ True positive + Σ True negative/Σ Total population | Positive predictive value (PPV), Precision = <templatestyles src="Sfrac/styles.css" />Σ True positive/Σ Test outcome positive | False omission rate (FOR) = <templatestyles src="Sfrac/styles.css" />Σ False negative/Σ Test outcome negative | Positive likelihood ratio (LR+) = <templatestyles src="Sfrac/styles.css" />TPR/FPR | Diagnostic odds ratio (DOR) = <templatestyles src="Sfrac/styles.css" />LR+/LR− | |
False discovery rate (FDR) = <templatestyles src="Sfrac/styles.css" />Σ False positive/Σ Test outcome positive | Negative predictive value (NPV) = <templatestyles src="Sfrac/styles.css" />Σ True negative/Σ Test outcome negative | Negative likelihood ratio (LR−) = <templatestyles src="Sfrac/styles.css" />FNR/TNR |
See also
- Phi coefficient
- F1 score
- Cramér's V, a similar measure of association between nominal variables.
- Cohen's kappa
References
<templatestyles src="Reflist/styles.css" />
Cite error: Invalid <references>
tag; parameter "group" is allowed only.
<references />
, or <references group="..." />