There are several statistical methods used to assess inter rater reliability, including:
Cohen's Kappa: A statistic that measures agreement between two raters, correcting for agreement that occurs by chance. Intraclass Correlation Coefficient (ICC): Used for continuous data and can assess the reliability of multiple raters. Fleiss' Kappa: An extension of Cohen's Kappa for assessing the reliability of multiple raters. Percent Agreement: The simplest measure, calculating the percentage of times raters agree. However, it does not account for chance agreement.