KappaAcc: A program for assessing the adequacy of kappa.

Behav Res Methods

Department of Psychology, Georgia State University, Atlanta, GA, 30303, USA.

Published: February 2023

Categorical cutpoints used to assess the adequacy of various statistics-like small, medium, and large for correlation coefficients of .10, .30, and .50 (Cohen, Cohen, J. (1988). Statistical power analysis for the behavioral sciences (2nd ed.). Lawrence Erlbaum Associates.)-are as useful as they are arbitrary, but not all statistics are suitable candidates for categorical cutpoints. One such is kappa, a statistic that gauges inter-observer agreement corrected for chance (Cohen Educational and Psychological Measurement, 20(1), 37-46, Cohen, Educational and Psychological Measurement 20:37-46, 1960). Depending on circumstances, a specific value of kappa may be judged adequate in one case but not in another. Thus, no one value of kappa can be regarded as universally acceptable and the question for investigators should be, are observers accurate enough, not is kappa big enough. A principled way to assess whether a specific value of kappa is adequate is to estimate observer accuracy-how accurate would simulated observers need to be to have generated a specific value of kappa obtained by actual observers, given specific circumstances. Estimating observer accuracy based on a kappa table the user provides is what KappaAcc, the program described here, does.

Download full-text PDF

Source
http://dx.doi.org/10.3758/s13428-022-01836-1DOI Listing

Publication Analysis

Top Keywords

specific kappa
12
kappaacc program
8
kappa
8
categorical cutpoints
8
cohen educational
8
educational psychological
8
psychological measurement
8
program assessing
4
assessing adequacy
4
adequacy kappa
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!