# Concordance Correlation Coefficient

**0-fold Cross-Validation**, and kindly contributed to R-bloggers]. (You can report issue about the content on this page here)

Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.

If we collect independent pairs of observations from some bivariate distribution, then how can we estimate the expected squared perpendicular distance of each such point in the 2D plane from the 45-degree line?

Assume that the random two-dimensional vector follows a bivariate distribution with mean , and covariance matrix with entries , and .

The squared perpendicular distance of the random point from the 45-degree line is

see the figure below.
Thus, the expected value of the squared perpendicular distance *times two* (for notational convenience) is given by,

To answer the question raised above, we can estimate the value of equation based on pairs of observations substituting the respective sample mean, sample variance, and covariance estimates for respectively.

## Defining the Concordance Correlation Coefficient

That’s great, but why should we spend any time thinking about the expected distance from the 45-degree line? What’s interesting about it?

Apart from delighting in the pure joy of doing mathematics and taking pleasure in the experience of mathematical beauty… :joy: :stuck_out_tongue_closed_eyes: … A measure of distance from the 45-degree line naturally quantifies the (dis)agreement between the two sets of observations. For example, we may have measured the same target entities using two different measurement instruments, and may want to know if and to what extent they agree.

Towards quantifying the extent of the (dis)agreement between two sets of observations it is natural to try to scale (or normalize) the quantity of equation to the range . However, it turns out that, rather than scaling to a range, it is customary to scale this quantity to the range from -1 to 1 as follows,

This expression, first introduced by (Lin, 1989), is known as the *Concordance Correlation Coefficient*, abbreviated as *CCC* hereafter.

## Concordance Correlation Coefficient vs. Pearson correlation coefficient

The scaling into the range from -1 to 1 may have been motivated by the fact that the Pearson correlation coefficient also falls within the range. In fact, analogous to how a Pearson correlation coefficient signifies perfect positive correlation, a CCC of 1 designates that the paired observations fall exactly on the line of perfect concordance (i.e., the 45-degree diagonal line).

Further aspects of the relationship to the Pearson correlation coefficient become visible if we rewrite the CCC further into the following set of equations.

where

From equations and we observe that:

- CCC always has the same sign as .
- CCC is 0 if and only if (with exception of cases when is undefined but CCC can still be computed via equation ).
- We can consider to be a measure of precision, i.e., how far each point deviates from the best fitting line. Thus, CCC combines (as a measure of precision) with an additional measure of accuracy denoted by in equations and , whereby:
- quantifies how far the best-fit line deviates from the 45-degree line.
- When the data lie exactly on the 45-degree line, then we have that C = 1. The further the observations deviate from the 45-degree line, the further C is from 1 and the closer it is to 0.
- In particular, from equation can be considered a measure of scale shift, and from equation a measure of location shift relative to the scale.

Now it turns out that the Pearson correlation coefficient has one major shortcoming when assessing reproducibility of measurements, such as when comparing two instruments that measure the same target entity.

:point_right: *Unlike CCC, is invariant to additive or multiplicative shifts by a constant value, referred to as location shift and scale shift respectively in the following set of figures:*

Looking at the above figures we see that the magnitude of the Pearson correlation coefficient does not change under location and scale shift (though the sign may flip). The CCC on the other hand quantifies the deviation from the 45-degree line, which is due to location and scale shifts in these examples, rather well.

This makes the CCC a better metric when we want to assess how well one measurement can reproduce another (i.e., how close the measurement pairs fall to the 45-degree line), while we would use if what we want is quantifying to what extent the measurement pairs can be described by a linear equation (with any intercept and slope).

The following figures show the same examples where both the and the coordinates are augmented with Gaussian noise (mean 0, standard deviation 15; the same realization of the random noise is used within each subfigure). We see that both and CCC move further away from the extreme values of , , and as noise is added.

## What’s CCC good for? Reproducibility?

As hinted above, you may want to compare two instruments that aim to measure the same target entity, or two assays that aim to measure the same analyte, or other quantitative measurement procedures or devices.
For example, one set of measurements may be obtained by what’s considered the “gold standard”, while the other set of measurements may be collected by a new instrument/assay/device that may be cheaper or in some other way preferable to the “gold standard” instrument/assay/device. Then one would wish to demonstrate that the collected two sets of measurements are equivalent.
(Lin, 1989) refers to this type of agreement or similarity between two sets of measurements as *reproducibility* of measurements.
The paper considers the following two illustrative examples:

(1) Can a “Portable $ave” machine (actual name withheld) reproduce a gold-standard machine in measuring total bilirubin in blood?

(2) Can an in-vitro assay for screening the toxicity of biomaterials reproduce from trial to trial?

And indeed this type of reproducibility assessment is a task where CCC has some clear advantages over the Pearson correlation coefficient, as seen in the figures above, as well as over some other approaches, as discussed in (Lin, 1989) in detail. A couple of shortcomings of common statistical approaches (when applied to the reproducibility assessment problem in question) are the following:

- The
*paired t-test*can reject a highly reproducible set of measurements when the variance is very small, while it fails to detect poor agreement in pairs of data when the means are equal. - The
*least squares approach*of testing intercept equal to 0 and slope equal to 1 fails to detect nearly perfect agreement when the residual errors are very small, while it has increasingly little chance to reject the null hypothesis of agreement the more the data are scattered.

I will end here. However, if you want to go deeper into the topic I invite you to check out the original paper by Lin for a more thorough discussion of the merits of the CCC as well as for its statistical properties. Moreover, since the publication of (Lin, 1989) there of course has been follow-up work, which I didn’t read (so, I may update this blog post in the future).

## References

- Lin, L. I. (1989). A concordance correlation coefficient to evaluate reproducibility.
*Biometrics*,*45*(1), 255–268. https://www.ncbi.nlm.nih.gov/pubmed/2720055

**leave a comment**for the author, please follow the link and comment on their blog:

**0-fold Cross-Validation**.

R-bloggers.com offers

**daily e-mail updates**about R news and tutorials about learning R and many other topics. Click here if you're looking to post or find an R/data-science job.

Want to share your content on R-bloggers? click here if you have a blog, or here if you don't.