Inter-Annotator Agreement: An Introduction to Cohen's Kappa Statistic | by Surge AI | Medium
What is Kappa and How Does It Measure Inter-rater Reliability?
PDF] Large sample standard errors of kappa and weighted kappa. | Semantic Scholar
kappa - Stata
StatHand - Calculating and interpreting a weighted kappa in SPSS - YouTube
Weighted Kappa in R: Best Reference - Datanovia
Inter-rater agreement
Cohen's Kappa Explained | Built In
PDF] Cohen's quadratically weighted kappa is higher than linearly weighted kappa for tridiagonal agreement tables | Semantic Scholar
Symmetry | Free Full-Text | An Empirical Comparative Assessment of Inter-Rater Agreement of Binary Outcomes and Multiple Raters
Using appropriate Kappa statistic in evaluating inter-rater reliability. Short communication on “Groundwater vulnerability and contamination risk mapping of semi-arid Totko river basin, India using GIS-based DRASTIC model and AHP techniques ...