Skip to main content icon/video/no-internet

Intercoder Reliability Techniques: Krippendorff’s Alpha

Krippendorff’s alpha coefficient is a statistical measure of the agreement among multiple replications of data-making processes. It emerged in the 1970s in content analyses projects whereby textual or visual units of analysis were categorized or valued by trained coders following written instructions. It has since been used in counseling and survey research where experts code open-ended interviews into analyzable terms, in medical laboratories where alternative methods of testing need to be compared, and in computational linguistics where the reliability of identifying syntactical or semantic parts of corpora is an issue.

Krippendorff’s alpha generalizes several known agreement statistics. It is applicable to any number of observers, coders, or judges, not just two; accepts any number of categories or values; can handle binary, nominal, ordinal, interval, and ratio metrics (levels of measurements); accepts incomplete or missing data; adjusts itself to small sample sizes; and is applicable not just to coding predefined units of analysis but also to unitizing or segmenting a continuum into units of various lengths. The virtue of a single coefficient with these variations is that the reliabilities it yields are comparable across a variety of data.

The cα-Coefficient for Assessing the Reliability of Coding Predefined Units

Reliability Data

Following Klaus Krippendorff’s Content Analysis: An Introduction to Its Methodology, 3rd Edition (2013, pp. 277–301), the canonical form of reliability data for one variable consists of a table of the categories or values c assigned to a set of nu units u, each by at least m ≥ 2 observers, coders, judges, or measuring instruments i or j.

Table 1 Reliability Data for Predefined Units

None

The General Form of Alpha

Agreement coefficients that aspire to be interpretable as indicators of reliability have to have two numerically distinct reference points at which reliability is perfect or absent, respectively. They are embedded in α’s definition:

α = 1 D o D e

where Do is the disagreement observed within the columns of the reliability data in Table 1, and De is the disagreement that is expected when reliability is absent, defined as the total lack of any relationship between the data and the phenomena of interest. Algebraically, when all values within the colums of Table 1 are the same but vary across them, Do = 0 and α = 1. When the observed disagreement is indistinguishable from what can be expected in the absence of any relationship to the phenomena being coded, Do = De and α = 0. α may become negative when disagreement is systematic, when observers agree to disagree, or when observers use different coding instructions. This condition has little to do with reliability. Reliable data that are worth considering should measure close to α = 1. So, the practical range of α’s values is

1 α 0

Coincidence Matrix Representations of Reliability Data

Coincidence matrices represent reliability data in terms of pairs of values c and k in the columns of Table 1. They do not preserve the references to the observers who assigned these values to units nor to the units in which they were found. They provide conceptually convenient tabulations.

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading