Kurtosis (from ( or ), meaning 'curved, arching') refers to the degree of tailedness in the probability distribution of a real-valued, random variable in probability theory and statistics. Similar to skewness, kurtosis provides insight into specific characteristics of a distribution. Various methods exist for quantifying kurtosis in theoretical distributions, and corresponding techniques allow estimation based on sample data from a population. It is important to note that different measures of kurtosis can yield varying interpretations.
The standard measure of a distribution's kurtosis, originating with Karl Pearson, is a scaled version of the fourth moment of the distribution. This number is related to the tails of the distribution, not its peak; hence, the sometimes-seen characterization of kurtosis as peakedness is incorrect. For this measure, higher kurtosis corresponds to greater extremity of deviations (or ), and not the configuration of data near the mean.
Excess kurtosis, typically compared to a value of 0, characterizes the tailedness of a distribution. A univariate normal distribution has an excess kurtosis of 0. Negative excess kurtosis indicates a platykurtic distribution, which does not necessarily have a flat top but produces fewer or less extreme outliers than the normal distribution. For instance, the uniform distribution (i.e., one that is uniformly finite over some bound and zero elsewhere) is platykurtic. On the other hand, positive excess kurtosis signifies a leptokurtic distribution. The Laplace distribution for example, has tails that decay more slowly than a normal one, resulting in more outliers. To simplify comparison with the normal distribution, excess kurtosis is calculated as Pearson's kurtosis minus 3. Some authors and software packages use kurtosis to refer specifically to excess kurtosis, but this article distinguishes between the two for clarity.
Alternative measures of kurtosis are: the L-kurtosis, which is a scaled version of the fourth L-moment; measures based on four population or sample quantiles. These are analogous to the alternative measures of skewness that are not based on ordinary moments.
The kurtosis is bounded below by the squared skewness plus 1: where is the third central moment. The lower bound is realized by the Bernoulli distribution. There is no upper limit to the kurtosis of a general probability distribution, and it may be infinite.
A reason why some authors favor the excess kurtosis is that cumulants are extensive. Formulas related to the extensive property are more naturally expressed in terms of the excess kurtosis. For example, let be independent random variables for which the fourth moment exists, and let be the random variable defined by the sum of the . The excess kurtosis of iswhere is the standard deviation of . In particular if all of the have the same variance, then this simplifies to
The reason not to subtract 3 is that the bare moment better generalizes to multivariate distributions, especially when independence is not assumed. The cokurtosis between pairs of variables is an order four tensor. For a bivariate normal distribution, the cokurtosis tensor has off-diagonal terms that are neither 0 nor 3 in general, so attempting to "correct" for an excess becomes confusing. It is true, however, that the joint cumulants of degree greater than two for any multivariate normal distribution are zero.
For two random variables, and , not necessarily independent, the kurtosis of the sum, , is Note that the fourth-power binomial coefficients (1, 4, 6, 4, 1) appear in the above equation.
Numerous misconceptions about kurtosis relate to notions of peakedness. One such misconception is that kurtosis measures both the peakedness of a distribution and the heaviness of its tail. Other incorrect interpretations include notions like lack of shoulders (where the shoulder refers vaguely to the area between the peak and the tail, or more specifically, the region about one standard deviation from the mean) or bimodality. Balanda and MacGillivray argue that the standard definition of kurtosis "poorly captures the kurtosis, peakedness, or tail weight of a distribution." Instead, they propose a vague definition of kurtosis as the location- and scale-free movement of probability mass from the distribution's shoulders into its center and tails.
Now by definition of the kurtosis , and by the well-known identity
The kurtosis can now be seen as a measure of the dispersion of around its expectation. Alternatively it can be seen to be a measure of the dispersion of around and . attains its minimal value in a symmetric two-point distribution. In terms of the original variable , the kurtosis is a measure of the dispersion of around the two values .
High values of arise where the probability mass is concentrated around the mean and the data-generating process produces occasional values far from the mean, or where the probability mass is concentrated in the tails of the distribution.
For any with positive definite, among all probability distributions on with mean and covariance , the normal distribution has the largest entropy.
Since mean and covariance are the first two moments, it is natural to consider extension to higher moments. In fact, by Lagrange multiplier method, for any prescribed first n moments, if there exists some probability distribution of form that has the prescribed moments (if it is feasible), then it is the maximal entropy distribution under the given constraints.
By serial expansion, so if a random variable has probability distribution , where is a normalization constant, then its kurtosis is .
All densities in this family are symmetric. The -th moment exists provided . For the kurtosis to exist, we require . Then the mean and skewness exist and are both identically zero. Setting makes the variance equal to unity. Then the only free parameter is , which controls the fourth moment (and cumulant) and hence the kurtosis. One can reparameterize with , where is the excess kurtosis as defined above. This yields a one-parameter leptokurtic family with zero mean, unit variance, zero skewness, and arbitrary non-negative excess kurtosis. The reparameterized density is
In the limit as , one obtains the density which is shown as the red curve in the images on the right.
In the other direction as one obtains the standard normal density as the limiting distribution, shown as the black curve.
In the images on the right, the blue curve represents the density with excess kurtosis of 2. The top image shows that leptokurtic densities in this family have a higher peak than the mesokurtic normal density, although this conclusion is only valid for this select family of distributions. The comparatively fatter tails of the leptokurtic densities are illustrated in the second image, which plots the natural logarithm of the Pearson type VII densities: the black curve is the logarithm of the standard normal density, which is a parabola. One can see that the normal density allocates little probability mass to the regions far from the mean (has thin tails), compared with the blue curve of the leptokurtic Pearson type VII density with excess kurtosis of 2. Between the blue curve and the black are other Pearson type VII densities with = 1, 1/2, 1/4, 1/8, and 1/16. The red curve again shows the upper limit of the Pearson type VII family, with (which, strictly speaking, means that the fourth moment does not exist). The red curve decreases the slowest as one moves outward from the origin (has fat tails).
Note that in these cases the platykurtic densities have bounded support, whereas the densities with positive or zero excess kurtosis are supported on the whole real line.
One cannot infer that high or low kurtosis distributions have the characteristics indicated by these examples. There exist platykurtic densities with infinite support, e.g., exponential power distributions with sufficiently large shape parameter b, and there exist leptokurtic densities with finite support. An example of the latter is a distribution that is uniform between −3 and −0.3, between −0.3 and 0.3, and between 0.3 and 3, with the same density in the (−3, −0.3) and (0.3, 3) intervals, but with 20 times more density in the (−0.3, 0.3) interval.
Also, one cannot infer from the graphs that higher kurtosis distributions are more peaked and that lower kurtosis distributions are more flat. There exist platykurtic densities with infinite peakedness; e.g., an equal mixture of the beta distribution with parameters 0.5 and 1 with its reflection about 0.0, and there exist leptokurtic densities that appear flat-topped; e.g., a mixture of distribution that is uniform between −1 and 1 with a T(4.0000001) Student's t-distribution, with mixing probabilities 0.999 and 0.001.
Graphs of the standardized versions of these distributions are given to the right.
This formula has the simpler representation,where the values are the standardized data values using the standard deviation defined using rather than in the denominator.
For example, suppose the data values are 0, 3, 4, 1, 2, 3, 0, 2, 1, 3, 2, 0, 2, 2, 3, 2, 5, 2, 3, 999.
Then the values are −0.239, −0.225, −0.221, −0.234, −0.230, −0.225, −0.239, −0.230, −0.234, −0.225, −0.230, −0.239, −0.230, −0.230, −0.225, −0.230, −0.216, −0.230, −0.225, 4.359
and the values are 0.003, 0.003, 0.002, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.003, 0.002, 0.003, 0.003, 360.976.
The average of these values is 18.05 and the excess kurtosis is thus . This example makes it clear that data near the middle or peak of the distribution do not contribute to the kurtosis statistic, hence kurtosis does not measure peakedness. It is simply a measure of the outlier, 999 in this example.
Unfortunately, in non-normal samples is itself generally biased.
Stated differently, under the assumption that the underlying random variable is normally distributed, it can be shown that .
D'Agostino's K-squared test is a goodness-of-fit normality test based on a combination of the sample skewness and sample kurtosis, as is the Jarque–Bera test for normality.
For non-normal samples, the variance of the sample variance depends on the kurtosis; for details, please see variance.
Pearson's definition of kurtosis is used as an indicator of intermittency in turbulence. It is also used in magnetic resonance imaging to quantify non-Gaussian diffusion.
A concrete example is the following lemma by He, Zhang, and Zhang: Assume a random variable has expectation , variance and kurtosis Assume we sample many independent copies. Then
This shows that with many samples, we will see one that is above the expectation with probability at least . In other words: If the kurtosis is large, there may be a lot of values either all below or above the mean.
|
|