Hi
With the Confusion Matrix below I get a Kappa value of 0,8023. Actually
it should be 1, because there are no wrong classifications.
I debugged and found out that it happens, because the samples variable
is two times incremented in the "putCount" method
(org.apache.mahout.classifier.ConfusionMatrix). Commenting out this
incrementation lead to the Kappa value of 1.
Is this a bug or do I something something?
=======================================================
Confusion Matrix
-------------------------------------------------------
a b <--Classified as
6 0 | 6 a = 1_YES
0 9 | 9 b = 2_NO
=======================================================
Statistics
-------------------------------------------------------
Kappa 0,8023
Accuracy 100%
Reliability 66,6667%
Reliability (standard deviation) 0,5774