mik07 wrote:
> 
> Hi,
> 
> this is more a general statistics question I think.
> 
> I am working on a system which automatically answers user questions (such
> systems are commonly called "Question Answering systems").
> I evaluated different versions of the same system on a publicly available
> test set.
> This set contains 500 question. Naturally, for each question the answer
> can be wrong or right, which is coded as "0" (wrong) or "1" (correct). By
> adding up all values, and dividing them by the number of questions in the
> test set (that's 500), one gets a measure for how well the system
> performs, commonly called accuracy.
> As mentioned I evaluated two different versions of the system, and
> received two different accuracy values. Now I want to know whether the
> difference is statistically significant. 
> 

?prop.test

-- 
View this message in context: 
http://www.nabble.com/Help-with-significance.-T-test--tp24699690p24701848.html
Sent from the R help mailing list archive at Nabble.com.

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to