From: Measuring agreement between healthcare survey instruments using mutual information
Data set | Measure | ||
---|---|---|---|
Odds ratio | Cohen’s kappa | Local mutual information | |
(Table 2 in Shulman et al. 1986) Clock exam and MMSEa | 15.6 | 0.493 | Agreement (0.422 in I agreement , -0.212 in I disagreement ) |
(Figure 1 in Russell et al. 2012) BDIb + CDRS_Rc and ICD-10d | ∞** | 0.693 | Agreement (0.18 in I agreement , -0.03 in I disagreement ) |
(Figure 1 in Russell et al. 2012) CDRS_R and ICD-10 | 0.311* | -0.015* | Inconclusive ( -0.018 in I agreement , 0.024 in I disagreement ) |
(Table 4 in Seago 2002) Score scheme comparison | 4 | 0.265 | Agreement ( 0.152 in I agreement , -0.107 in I disagreement ) |