Bayesian inference of dependent kappa for binary ratings

Stat Med. 2021 Nov 20;40(26):5947-5960. doi: 10.1002/sim.9165. Epub 2021 Sep 20.

Abstract

In medical and social science research, reliability of testing methods measured through inter- and intraobserver agreement is critical in disease diagnosis. Often comparison of agreement across multiple testing methods is sought in situations where testing is carried out on the same experimental units rendering the outcomes to be correlated. In this article, we first developed a Bayesian method for comparing dependent agreement measures under a grouped data setting. Simulation studies showed that the proposed methodology outperforms the competing methods in terms of power, while maintaining a decent type I error rate. We further developed a Bayesian joint model for comparing dependent agreement measures adjusting for subject and rater-level heterogeneity. Simulation studies indicate that our model outperforms a competing method that is used in this context. The developed methodology was implemented on a key measure on a dichotomous rating scale from a study with six raters evaluating three classification methods for chest radiographs for pneumoconiosis developed by the International Labor Office.

Keywords: Bayesian inference; correlated kappa; covariate adjustment; grouped data; test of homogeneity.

Publication types

  • Research Support, Non-U.S. Gov't
  • Research Support, U.S. Gov't, P.H.S.

MeSH terms

  • Bayes Theorem*
  • Computer Simulation
  • Humans
  • Observer Variation
  • Reproducibility of Results