Hi,

This is good survey paper that can be found regard to Anomaly detection
[1], According to your need; it seems you will no need to go through whole
the survey papers. But few sub topics will be very useful for you. This
paper will be useful for your work.

[1] Varun Chandola, Arindam Banerjee, and Vipin Kumar. 2009. Anomaly
detection: A survey. ACM Comput. Surv. 41, 3, Article 15 (July 2009), 58
pages. DOI=10.1145/1541880.1541882
<http://www.researchgate.net/profile/Vipin_Kumar26/publication/220565847_Anomaly_detection_A_survey/links/0deec5161f0ca7302a000000.pdf>
[Cited by 2458]

On Wed, Sep 16, 2015 at 3:35 PM, Ashen Weerathunga <as...@wso2.com> wrote:

> Hi all,
>
> I am currently doing the integration of anomaly detection feature for ML.
> I have a problem of choosing the best accuracy measure for the model. I can
> get the confusion matrix which consists of true positives, true negatives,
> false positives and false negatives. There are few different measures such
> as sensitivity, accuracy, F1 score, etc. So what will be the best measure
> to give as the model accuracy for anomaly detection model.
>
> [1] <https://en.wikipedia.org/wiki/Sensitivity_and_specificity>Some
> details about those measures.
>
> Terminology and derivations
> from a confusion matrix <https://en.wikipedia.org/wiki/Confusion_matrix> true
> positive (TP)eqv. with hittrue negative (TN)eqv. with correct rejectionfalse
> positive (FP)eqv. with false alarm
> <https://en.wikipedia.org/wiki/False_alarm>, Type I error
> <https://en.wikipedia.org/wiki/Type_I_error>false negative (FN)eqv. with
> miss, Type II error <https://en.wikipedia.org/wiki/Type_II_error>
> ------------------------------
> sensitivity <https://en.wikipedia.org/wiki/Sensitivity_%28test%29> or
> true positive rate (TPR)eqv. with hit rate
> <https://en.wikipedia.org/wiki/Hit_rate>, recall
> <https://en.wikipedia.org/wiki/Information_retrieval#Recall>[image:
> \mathit{TPR} = \mathit{TP} / P = \mathit{TP} / (\mathit{TP}+\mathit{FN})]
> specificity <https://en.wikipedia.org/wiki/Specificity_%28tests%29> (SPC)
> or true negative rate[image: \mathit{SPC} = \mathit{TN} / N = \mathit{TN}
> / (\mathit{TN}+\mathit{FP})]precision
> <https://en.wikipedia.org/wiki/Information_retrieval#Precision> or positive
> predictive value <https://en.wikipedia.org/wiki/Positive_predictive_value>
> (PPV)[image: \mathit{PPV} = \mathit{TP} / (\mathit{TP} + \mathit{FP})]negative
> predictive value <https://en.wikipedia.org/wiki/Negative_predictive_value>
> (NPV)[image: \mathit{NPV} = \mathit{TN} / (\mathit{TN} + \mathit{FN})]
> fall-out <https://en.wikipedia.org/wiki/Information_retrieval#Fall-out>
> or false positive rate <https://en.wikipedia.org/wiki/False_positive_rate>
> (FPR)[image: \mathit{FPR} = \mathit{FP} / N = \mathit{FP} / (\mathit{FP}
> + \mathit{TN}) = 1-\mathit{SPC}]false negative rate
> <https://en.wikipedia.org/wiki/False_negative_rate> (FNR)[image:
> \mathit{FNR} = \mathit{FN} / (\mathit{TP} + \mathit{FN}) = 
> 1-\mathit{TPR}]false
> discovery rate <https://en.wikipedia.org/wiki/False_discovery_rate> 
> (FDR)[image:
> \mathit{FDR} = \mathit{FP} / (\mathit{TP} + \mathit{FP}) = 1 - \mathit{PPV}]
> ------------------------------
> accuracy <https://en.wikipedia.org/wiki/Accuracy> (ACC)[image:
> \mathit{ACC} = (\mathit{TP} + \mathit{TN}) / (\mathit{TP} + \mathit{FP} +
> \mathit{FN} + \mathit{TN})]F1 score
> <https://en.wikipedia.org/wiki/F1_score>is the harmonic mean
> <https://en.wikipedia.org/wiki/Harmonic_mean#Harmonic_mean_of_two_numbers>
> of precision
> <https://en.wikipedia.org/wiki/Information_retrieval#Precision> and
> sensitivity <https://en.wikipedia.org/wiki/Sensitivity_%28test%29>[image:
> \mathit{F1} = 2 \mathit{TP} / (2 \mathit{TP} + \mathit{FP} + 
> \mathit{FN})]Matthews
> correlation coefficient
> <https://en.wikipedia.org/wiki/Matthews_correlation_coefficient> (MCC)[image:
> \frac{ \mathit{TP} \times \mathit{TN} - \mathit{FP} \times \mathit{FN} }
> {\sqrt{ (\mathit{TP}+\mathit{FP}) ( \mathit{TP} + \mathit{FN} ) (
> \mathit{TN} + \mathit{FP} ) ( \mathit{TN} + \mathit{FN} ) } 
> }]Informedness[image:
> \mathit{TPR} + \mathit{SPC} - 1]Markedness
> <https://en.wikipedia.org/wiki/Markedness>[image: \mathit{PPV} +
> \mathit{NPV} - 1]
>
> *Sources: Fawcett (2006) and Powers (2011).*[1]
> <https://en.wikipedia.org/wiki/Sensitivity_and_specificity#cite_note-Fawcett2006-1>
> [2]
> <https://en.wikipedia.org/wiki/Sensitivity_and_specificity#cite_note-Powers2011-2>
>
> Thanks and Regards,
> Ashen
> --
> *Ashen Weerathunga*
> Software Engineer - Intern
> WSO2 Inc.: http://wso2.com
> lean.enterprise.middleware
>
> Email: as...@wso2.com
> Mobile: +94 716042995 <94716042995>
> LinkedIn:
> *http://lk.linkedin.com/in/ashenweerathunga
> <http://lk.linkedin.com/in/ashenweerathunga>*
>
> _______________________________________________
> Dev mailing list
> Dev@wso2.org
> http://wso2.org/cgi-bin/mailman/listinfo/dev
>
>


-- 
Cheers,
Madhuka Udantha
http://madhukaudantha.blogspot.com
_______________________________________________
Dev mailing list
Dev@wso2.org
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to