Detecting systematic bias between two raters.

Clin Exp Pharmacol Physiol

The University of Melbourne, Parkville, 563 Canning Street, Carlton North, Victoria 3054, Australia.

Published: September 2004

1. I recently reviewed, inter alia, methods for comparing two raters who make judgements on an ordered categorical scale, directed principally at the kappa statistic and its weaknesses. 2. The main weakness of the kappa statistic is that it fails to detect the all-important feature of systematic bias between raters. 3. I described various methods for detecting bias between two raters. These included a modified McNemar test and the single binomial test. Others that have been suggested are the symmetry of disagreement index (SD) and the marginal homogeneity test. 4. I now realize that none of the above four tests for bias is satisfactory, because all ignore the extent of agreement. 5. The bias index (BI) does take into account the extent of agreement, but its inventors did not propose how BI could be evaluated. I now describe a method for doing this.

Download full-text PDF

Source
http://dx.doi.org/10.1111/j.1440-1681.2004.03960.xDOI Listing

Publication Analysis

Top Keywords

bias raters
12
systematic bias
8
kappa statistic
8
extent agreement
8
bias
5
detecting systematic
4
raters
4
raters reviewed
4
reviewed inter
4
inter alia
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!