HOW reliable are change‐corrected measures of agreement?

Abstract
Chance‐corrected measures of agreement are prone to exhibit paradoxical and counter‐intuitive results when used as measures of reliability. It is demonstrated that these problems arise with Cohen's kappa as well as with Aickin's alpha. They are the consequence of an analogy to Simpson's paradox in mixed populations. It is further shown that chance‐corrected measures of agreement may yield misleading values for binary ratings. It is concluded that improvements in the design and the analysis of reliability studies are a pre‐requisite for valid and pertinent results.

This publication has 31 references indexed in Scilit: