agree_coef(
wide = TRUE,
col.names = NULL,
measure,
item,
id,
data,
weighted = FALSE,
conf.level = 0.95
)
Logical value (TRUE or FALSE) indicating if data is in a "wide" format. Default is TRUE.
If wide is equal to TRUE then col.names is a list of the column names containing the measurements for reliability analysis.
Name of column containing the measurement of interest.
Name of column containing the items. If this is an inter-rater reliability study then this would indicate the rater (e.g., rater1, rater2, rater3, etc).
Column with subject identifier.
Data frame with all data.
Logical value (TRUE or FALSE) indicating whether to weight the responses. If TRUE (default is FALSE) then quadratic weights are utilized. This option should be set to TRUE for ordinal or continuous responses.
the confidence level required. Default is 95%.
Returns single data frame of inter-rater reliability coefficients.
Gwet, K.L. (2014, ISBN:978-0970806284). “Handbook of Inter-Rater Reliability,” 4th Edition. Advanced Analytics, LLC. Gwet, K. L. (2008). “Computing inter-rater reliability and its variance in the presence of high agreement," British Journal of Mathematical and Statistical Psychology, 61, 29-48.
data('reps')
agree_coef(data = reps, wide = TRUE, col.names = c("x","y"), weighted = TRUE)
#> est se lower.ci upper.ci
#> Percent Agreement 0.9131459 0.07673687 0.75253380 1.0000000
#> Gwet's AC2 0.5073626 0.18354840 0.12319136 0.8915338
#> Fleiss' Kappa 0.5255860 0.17484095 0.15963973 0.8915324
#> Kririppendorff's Alpha 0.4771011 0.19501219 0.06566137 0.8885409