Confusion Matrix


A confusion matrix (Kohavi and Provost, 1998) contains information aboutactual and predicted classifications done by a classification system. Performance of such systems is commonly evaluated using the data in the matrix. The following table shows the confusion matrix for a two class classifier.

The entries in the confusion matrix have the following meaning in the context of our study:

  • a is the number of correct predictions that an instance is negative,
  • b is the number of incorrect predictions that an instance is positive,
  • c is the number of incorrect of predictions that an instance negative, and
  • d is the number of correct predictions that an instance is positive.
 Predicted
NegativePositive
ActualNegativeab
Positivecd

Several standard terms have been defined for the 2 class matrix:

  • The accuracy (AC) is the proportion of the total number of predictions that were correct. It is determined using the equation:

cm1
[1]

  • The recall or true positive rate (TP) is the proportion of positive cases that were correctly identified, as calculated using the equation:

cm2
[2]

  • The false positive rate (FP) is the proportion of negatives cases that were incorrectly classified as positive, as calculated usingthe equation:

cm3
[3]

  • The true negative rate (TN) is defined as the proportion of negatives cases that were classified correctly, as calculated using the equation:

cm4
[4]

  • The false negative rate (FN) is the proportion of positives cases that were incorrectly classified as negative, as calculated using the equation:

cm5
[5]

  • Finally, precision (P) is the proportion of the predicted positive cases that were correct, as calculated using the equation:

cm6
[6]

The accuracy determined using equation 1 may not be an adequate performance measure when the number of negative cases is much greater than the number of positive cases (Kubat et al., 1998). Suppose there are 1000 cases, 995 of which are negative cases and 5 of which are positive cases. If the systemclassifies them all as negative, the accuracy would be 99.5%, even though the classifier missed all positive cases. Other performance measures accountfor this by including TP in a product: for example, geometric mean (g-mean) (Kubat et al., 1998), as defined in equations7 and 8, and F-Measure (Lewis and Gale, 1994), as defined in equation 9.

cm7
[7]

cm8
[8]

cm9
[9]

In equation 9, b has a value from 0 to infinityand is used to control the weight assigned to TP and P. Anyclassifier evaluated using equations 7, 8 or 9 will have a measure valueof 0, if all positive cases are classified incorrectly.

Another way to examine the performance of classifiers is to use a ROCgraph, described on the next page.

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值