MedCalc

# Intraclass correlation coefficient

 Command: StatisticsAgreement & responsivenessIntraclass correlation coefficient

## Description

The Intraclass Correlation Coefficient (ICC) is a measure of the reliability of measurements or ratings.

For the purpose of assessing inter-rater reliability and the ICC, two or preferably more raters rate a number of study subjects.

• A distinction is made between two study models: (1) each subject is rated by a different and random selection of a pool of raters, and (2) each subject is rated by the same raters.
• In the first model, the ICC is always a measure for Absolute agreement; in the second model a choice can be made between two types: Consistency when systematic differences between raters are irrelevant, and Absolute agreement, when systematic differences are relevant.
For example: the paired ratings (2,4), (4,6) and (6,8) are in perfect agreement, with a consistency of 1.0, but with an absolute agreement of 0.6667.

## How to enter data

In this example (taken from Shrout PE & Fleiss JL, 1979) data are available for 4 raters on 6 subjects. The data for each subject are entered in the 4 columns.

If not all subjects are rated by the same 4 raters, the data are still entered in 4 columns, the order of which then being unimportant.

## Required input

• Measurements: variables that contain the measurements of the different raters.
• Filter: an optional filter to include only a selected subgroup of cases.
• Options
• Model
• Raters for each subject were selected at random: the raters were not the same for all subjects, a random selection or raters rated each subject.
• The same raters for all subjects: all subjects were rated by the same raters.
• Type
• Consistency: systematic differences between raters are irrelevant.
• Absolute agreement: systematic differences are relevant

## Results

The Intraclass correlation coefficient table reports two coefficients with their respective 95% Confidence Interval.

• Single measures: this ICC is an index for the reliability of the ratings for one, typical, single rater.
• Average measures: this ICC is an index for the reliability of different raters averaged together. This ICC is always higher than the Single measures ICC.

## Literature

• McGraw KO, Wong SP (1996) Forming inferences about some intraclass correlation coefficients. Psychological Methods 1:30-46. (Correction: 1:390).
• Shrout PE, Fleiss JL (1979) Intraclass correlations: uses in assessing rater reliability. Psychological Bulletin 86:420-428.