How to calculate Fleiss’ Kappa in Excel?

Fleiss’ Kappa is a statistic used to measure inter-rater agreement for multiple raters. It can be calculated in Excel using the KAPPA.DIST function. This function takes four arguments: the number of raters, the number of categories, and two matrices containing the number of agreements and disagreements between the raters. The function returns the Kappa statistic, which can be interpreted as a measure of agreement between the raters.


Fleiss’ Kappa is a way to measure the degree of agreement between three or more raters when the raters are assigning categorical ratings to a set of items.

Fleiss’ Kappa ranges from 0 to 1 where:

  • indicates no agreement at all among the raters.
  • indicates perfect inter-rater agreement.

This tutorial provides an example of how to calculate Fleiss’ Kappa in Excel.

Example: Fleiss’ Kappa in Excel

Suppose 14 individuals rate 10 different products on a scale of Poor to Excellent.

The following screenshot displays the total ratings that each product received:

The following screenshot shows how to calculate Fleiss’ Kappa for this data in Excel:

Fleiss' Kappa calculation in Excel

The trickiest calculations in this screenshot are in column J. The formula used for these calculations is shown in the text box near the top of the screen.

Note that the Fleiss’ Kappa in this example turns out to be 0.2099. The actual formula used to calculate this value in cell C18 is:

Fleiss’ Kappa = (0.37802 – 0.2128) / (1 –  0.2128) = 0.2099.

Although there is no formal way to interpret Fleiss’ Kappa, the following values show how to interpret Cohen’s Kappa, which is used to assess the level of inter-rater agreement between just two raters:

  • < 0.20 | Poor
  • .21 – .40 | Fair
  • .41 – .60 | Moderate
  • .61 – .80 | Good
  • .81 – 1 | Very Good

Based on these values, Fleiss’ Kappa of 0.2099 in our example would be interpreted as a “fair” level of inter-rater agreement.

x