In statistics, the Fisher transformation (or Fisher z-transformation) of a Pearson correlation coefficient is its inverse hyperbolic tangent (artanh).
When the sample correlation coefficient r is near 1 or -1, its distribution is highly skewed, which makes it difficult to estimate confidence intervals and apply tests of significance for the population correlation coefficient ρ.[1][2][3]
The Fisher transformation solves this problem by yielding a variable whose distribution is approximately normally distributed, with a variance that is stable over different values of r.
Here stands for the covariance between the variables and and stands for the standard deviation of the respective variable. Fisher's z-transformation of r is defined as
where N is the sample size, and ρ is the true correlation coefficient.
This transformation, and its inverse
can be used to construct a large-sample confidence interval for r using standard normal theory and derivations. See also application to partial correlation.
Hotelling gives a concise derivation of the Fisher transformation.[4]
To derive the Fisher transformation, one starts by considering an arbitrary increasing, twice-differentiable function of , say . Finding the first term in the large- expansion of the corresponding skewness results[5] in
Setting and solving the corresponding differential equation for yields the inverse hyperbolic tangent function.
Similarly expanding the mean m and variance v of , one gets
m =
and
v =
respectively.
The extra terms are not part of the usual Fisher transformation. For large values of and small values of they represent a large improvement of accuracy at minimal cost, although they greatly complicate the computation of the inverse – a closed-form expression is not available. The near-constant variance of the transformation is the result of removing its skewness – the actual improvement is achieved by the latter, not by the extra terms. Including the extra terms, i.e., computing (z-m)/v1/2, yields:
The application of Fisher's transformation can be enhanced using a software calculator as shown in the figure. Assuming that the r-squared value found is 0.80, that there are 30 data [clarification needed], and accepting a 90% confidence interval, the r-squared value in another random sample from the same population may range from 0.656 to 0.888. When r-squared is outside this range, the population is considered to be different.
Discussion
The Fisher transformation is an approximate variance-stabilizing transformation for r when X and Y follow a bivariate normal distribution. This means that the variance of z is approximately constant for all values of the population correlation coefficient ρ. Without the Fisher transformation, the variance of r grows smaller as |ρ| gets closer to 1. Since the Fisher transformation is approximately the identity function when |r| < 1/2, it is sometimes useful to remember that the variance of r is well approximated by 1/N as long as |ρ| is not too large and N is not too small. This is related to the fact that the asymptotic variance of r is 1 for bivariate normal data.
The behavior of this transform has been extensively studied since Fisher introduced it in 1915. Fisher himself found the exact distribution of z for data from a bivariate normal distribution in 1921; Gayen in 1951[8]
determined the exact distribution of z for data from a bivariate Type A Edgeworth distribution. Hotelling in 1953 calculated the Taylor series expressions for the moments of z and several related statistics[9] and Hawkins in 1989 discovered the asymptotic distribution of z for data from a distribution with bounded fourth moments.[10]
An alternative to the Fisher transformation is to use the exact confidence distribution density for ρ given by[11][12]
where is the Gaussian hypergeometric function and .
^Fisher, R. A. (1915). "Frequency distribution of the values of the correlation coefficient in samples of an indefinitely large population". Biometrika. 10 (4): 507–521. doi:10.2307/2331838. hdl:2440/15166. JSTOR2331838.
^Gayen, A. K. (1951). "The Frequency Distribution of the Product-Moment Correlation Coefficient in Random Samples of Any Size Drawn from Non-Normal Universes". Biometrika. 38 (1/2): 219–247. doi:10.1093/biomet/38.1-2.219. JSTOR2332329.
^Hotelling, H (1953). "New light on the correlation coefficient and its transforms". Journal of the Royal Statistical Society, Series B. 15 (2): 193–225. JSTOR2983768.