Moment (mathematics)

From Wikipedia, the free encyclopedia
(Redirected from First moment)
Jump to navigation Jump to search

Template:Short description Script error: No such module "about".

Moments of a function in mathematics are certain quantitative measures related to the shape of the function's graph. For example, if the function represents mass density, then the zeroth moment is the total mass, the first moment (normalized by total mass) is the center of mass, and the second moment is the moment of inertia. If the function is a probability distribution, then the first moment is the expected value, the second central moment is the variance, the third standardized moment is the skewness, and the fourth standardized moment is the kurtosis.

For a distribution of mass or probability on a bounded interval, the collection of all the moments (of all orders, from 0Script error: No such module "Check for unknown parameters". to Script error: No such module "Check for unknown parameters".) uniquely determines the distribution (Hausdorff moment problem). The same is not true on unbounded intervals (Hamburger moment problem).

In the mid-nineteenth century, Pafnuty Chebyshev became the first person to think systematically in terms of the moments of random variables.[1]

Significance of the moments

The Template:Mvarth raw moment (i.e., moment about zero) of a random variable X with density function f(x) is defined by[2]μ'n=Xn=def{ixinf(xi),discrete distributionxnf(x)dx,continuous distributionThe Template:Mvarth moment of a real-valued continuous random variable with density function f(x) about a value c is the integralμn=(xc)nf(x)dx.

It is possible to define moments for random variables in a more general fashion than moments for real-valued functions – see moments in metric spaces. The moment of a function, without further explanation, usually refers to the above expression with c=0. For the second and higher moments, the central moment (moments about the mean, with c being the mean) are usually used rather than the moments about zero, because they provide clearer information about the distribution's shape.

Other moments may also be defined. For example, the Template:Mvarth inverse moment about zero is E[Xn] and the Template:Mvarth logarithmic moment about zero is E[lnn(X)].

The Template:Mvarth moment about zero of a probability density function f(x) is the expected value of Xn and is called a raw moment or crude moment.[3] The moments about its mean μ are called central moments; these describe the shape of the function, independently of translation.

If f is a probability density function, then the value of the integral above is called the Template:Mvarth moment of the probability distribution. More generally, if F is a cumulative probability distribution function of any probability distribution, which may not have a density function, then the Template:Mvarth moment of the probability distribution is given by the Riemann–Stieltjes integralμ'n=E[Xn]=xndF(x)where X is a random variable that has this cumulative distribution F, and EScript error: No such module "Check for unknown parameters". is the expectation operator or mean. WhenE[|Xn|]=|xn|dF(x)=the moment is said not to exist. If the Template:Mvarth moment about any point exists, so does the (n − 1)Script error: No such module "Check for unknown parameters".th moment (and thus, all lower-order moments) about every point. The zeroth moment of any probability density function is 1Script error: No such module "Check for unknown parameters"., since the area under any probability density function must be equal to one.

Significance of moments (raw, central, standardised) and cumulants (raw, normalised), in connection with named properties of distributions
Moment
ordinal
Moment Cumulant
Raw Central Standardized Raw Normalized
1 Mean 0 0 Mean
2 Variance 1 Variance 1
3 Skewness Skewness
4 (Non-excess or historical) kurtosis Excess kurtosis
5 Hyperskewness
6 Hypertailedness
7+

Standardized moments

Script error: No such module "Labelled list hatnote". The normalised Template:Mvarth central moment or standardised moment is the Template:Mvarth central moment divided by Template:Mvar; the normalised Template:Mvarth central moment of the random variable Template:Mvar is μnσn=E[(Xμ)n]σn=E[(Xμ)n]E[(Xμ)2]n2.

These normalised central moments are dimensionless quantities, which represent the distribution independently of any linear change of scale.

Notable moments

Mean

Script error: No such module "Labelled list hatnote". The first raw moment is the mean, usually denoted μE[X].

Variance

Script error: No such module "Labelled list hatnote". The second central moment is the variance. The positive square root of the variance is the standard deviation σ(E[(xμ)2])12.

Skewness

Script error: No such module "Labelled list hatnote". The third central moment is the measure of the lopsidedness of the distribution; any symmetric distribution will have a third central moment, if defined, of zero. The normalised third central moment is called the skewness, often Template:Mvar. A distribution that is skewed to the left (the tail of the distribution is longer on the left) will have a negative skewness. A distribution that is skewed to the right (the tail of the distribution is longer on the right), will have a positive skewness.

For distributions that are not too different from the normal distribution, the median will be somewhere near μγσ/6Script error: No such module "Check for unknown parameters".; the mode about μγσ/2Script error: No such module "Check for unknown parameters"..

Kurtosis

Script error: No such module "Labelled list hatnote".

The fourth central moment is a measure of the heaviness of the tail of the distribution. Since it is the expectation of a fourth power, the fourth central moment, where defined, is always nonnegative; and except for a point distribution, it is always strictly positive. The fourth central moment of a normal distribution is 3σ4Script error: No such module "Check for unknown parameters"..

The kurtosis Template:Mvar is defined to be the standardized fourth central moment. (Equivalently, as in the next section, excess kurtosis is the fourth cumulant divided by the square of the second cumulant.)[4][5] If a distribution has heavy tails, the kurtosis will be high (sometimes called leptokurtic); conversely, light-tailed distributions (for example, bounded distributions such as the uniform) have low kurtosis (sometimes called platykurtic).

The kurtosis can be positive without limit, but Template:Mvar must be greater than or equal to γ2 + 1Script error: No such module "Check for unknown parameters".; equality only holds for binary distributions. For unbounded skew distributions not too far from normal, Template:Mvar tends to be somewhere in the area of γ2Script error: No such module "Check for unknown parameters". and 2γ2Script error: No such module "Check for unknown parameters"..

The inequality can be proven by consideringE[(T2aT1)2]where T = (Xμ)/σScript error: No such module "Check for unknown parameters".. This is the expectation of a square, so it is non-negative for all a; however it is also a quadratic polynomial in a. Its discriminant must be non-positive, which gives the required relationship.

Higher moments

High-order moments are moments beyond 4th-order moments.

As with variance, skewness, and kurtosis, these are higher-order statistics, involving non-linear combinations of the data, and can be used for description or estimation of further shape parameters. The higher the moment, the harder it is to estimate, in the sense that larger samples are required in order to obtain estimates of similar quality. This is due to the excess degrees of freedom consumed by the higher orders. Further, they can be subtle to interpret, often being most easily understood in terms of lower order moments – compare the higher-order derivatives of jerk and jounce in physics. For example, just as the 4th-order moment (kurtosis) can be interpreted as "relative importance of tails as compared to shoulders in contribution to dispersion" (for a given amount of dispersion, higher kurtosis corresponds to thicker tails, while lower kurtosis corresponds to broader shoulders), the 5th-order moment can be interpreted as measuring "relative importance of tails as compared to center (mode and shoulders) in contribution to skewness" (for a given amount of skewness, higher 5th moment corresponds to higher skewness in the tail portions and little skewness of mode, while lower 5th moment corresponds to more skewness in shoulders).

Mixed moments

Mixed moments are moments involving multiple variables.

The value E[Xk] is called the moment of order k (moments are also defined for non-integral k). The moments of the joint distribution of random variables X1...Xn are defined similarly. For any integers ki0, the mathematical expectation E[X1k1Xnkn] is called a mixed moment of order k (where k=k1+...+kn), and E[(X1E[X1])k1(XnE[Xn])kn] is called a central mixed moment of order k. The mixed moment E[(X1E[X1])(X2E[X2])] is called the covariance and is one of the basic characteristics of dependency between random variables.

Some examples are covariance, coskewness and cokurtosis. While there is a unique covariance, there are multiple co-skewnesses and co-kurtoses.

Properties of moments

Transformation of center

Since (xb)n=(xa+ab)n=i=0n(ni)(xa)i(ab)ni where (ni) is the binomial coefficient, it follows that the moments about b can be calculated from the moments about a by: E[(xb)n]=i=0n(ni)E[(xa)i](ab)ni.

Moment of a convolution of function

Script error: No such module "Labelled list hatnote". The raw moment of a convolution h(t)=(f*g)(t)=f(τ)g(tτ)dτ reads μn[h]=i=0n(ni)μi[f]μni[g] where μn[] denotes the nth moment of the function given in the brackets. This identity follows by the convolution theorem for moment generating function and applying the chain rule for differentiating a product.

Cumulants

Script error: No such module "Labelled list hatnote".

The first raw moment and the second and third unnormalized central moments are additive in the sense that if X and Y are independent random variables then m1(X+Y)=m1(X)+m1(Y)Var(X+Y)=Var(X)+Var(Y)μ3(X+Y)=μ3(X)+μ3(Y)

(These can also hold for variables that satisfy weaker conditions than independence. The first always holds; if the second holds, the variables are called uncorrelated).

These are the first three cumulants and all cumulants share this additivity property.

Sample moments

For all k, the Template:Mvarth raw moment of a population can be estimated using the Template:Mvarth raw sample moment 1ni=1nXik applied to a sample X1, ..., XnScript error: No such module "Check for unknown parameters". drawn from the population.

It can be shown that the expected value of the raw sample moment is equal to the Template:Mvarth raw moment of the population, if that moment exists, for any sample size Template:Mvar. It is thus an unbiased estimator. This contrasts with the situation for central moments, whose computation uses up a degree of freedom by using the sample mean. So for example an unbiased estimate of the population variance (the second central moment) is given by 1n1i=1n(XiX¯)2 in which the previous denominator Template:Mvar has been replaced by the degrees of freedom n − 1Script error: No such module "Check for unknown parameters"., and in which X¯ refers to the sample mean. This estimate of the population moment is greater than the unadjusted observed sample moment by a factor of nn1, and it is referred to as the "adjusted sample variance" or sometimes simply the "sample variance".

Problem of moments

Script error: No such module "Labelled list hatnote". Problems of determining a probability distribution from its sequence of moments are called problem of moments. Such problems were first discussed by P.L. Chebyshev (1874)[6] in connection with research on limit theorems. In order that the probability distribution of a random variable X be uniquely defined by its moments αk=E[Xk] it is sufficient, for example, that Carleman's condition be satisfied: k=11α2k1/2k= A similar result even holds for moments of random vectors. The problem of moments seeks characterizations of sequences μn:n=1,2,3, that are sequences of moments of some function fScript error: No such module "Check for unknown parameters"., all moments αk(n) of which are finite, and for each integer k1 let αk(n)αk,n, where αk is finite. Then there is a sequence μn' that weakly converges to a distribution function μ having αk as its moments. If the moments determine μ uniquely, then the sequence μn' weakly converges to μ.

Partial moments

Partial moments are sometimes referred to as "one-sided moments". The Template:Mvarth order lower and upper partial moments with respect to a reference point r may be expressed as μn(r)=r(rx)nf(x)dx, μn+(r)=r(xr)nf(x)dx.

If the integral function does not converge, the partial moment does not exist.

Partial moments are normalized by being raised to the power 1/nScript error: No such module "Check for unknown parameters".. The upside potential ratio may be expressed as a ratio of a first-order upper partial moment to a normalized second-order lower partial moment.

Central moments in metric spaces

Let (M, d)Script error: No such module "Check for unknown parameters". be a metric space, and let B(M)Script error: No such module "Check for unknown parameters". be the [[Borel sigma algebra|Borel Template:Mvar-algebra]] on MScript error: No such module "Check for unknown parameters"., the [[sigma algebra|Template:Mvar-algebra]] generated by the dScript error: No such module "Check for unknown parameters".-open subsets of MScript error: No such module "Check for unknown parameters".. (For technical reasons, it is also convenient to assume that MScript error: No such module "Check for unknown parameters". is a separable space with respect to the metric dScript error: No such module "Check for unknown parameters"..) Let 1 ≤ p ≤ ∞Script error: No such module "Check for unknown parameters"..

The Template:Mvarth central moment of a measure Template:Mvar on the measurable space (M, B(M))Script error: No such module "Check for unknown parameters". about a given point x0MScript error: No such module "Check for unknown parameters". is defined to be Md(x,x0)pdμ(x).

μScript error: No such module "Check for unknown parameters". is said to have finite Template:Mvarth central moment if the Template:Mvarth central moment of Template:Mvar about x0Script error: No such module "Check for unknown parameters". is finite for some x0MScript error: No such module "Check for unknown parameters"..

This terminology for measures carries over to random variables in the usual way: if (Ω, Σ, P)Script error: No such module "Check for unknown parameters". is a probability space and X : Ω → MScript error: No such module "Check for unknown parameters". is a random variable, then the Template:Mvarth central moment of XScript error: No such module "Check for unknown parameters". about x0MScript error: No such module "Check for unknown parameters". is defined to be Md(x,x0)pd(X*(𝐏))(x)=Ωd(X(ω),x0)pd𝐏(ω)=𝐄[d(X,x0)p], and X has finite Template:Mvarth central moment if the Template:Mvarth central moment of XScript error: No such module "Check for unknown parameters". about x0Script error: No such module "Check for unknown parameters". is finite for some x0MScript error: No such module "Check for unknown parameters"..

See also

<templatestyles src="Div col/styles.css"/>

References

<templatestyles src="Reflist/styles.css" />

  1. Script error: No such module "Citation/CS1".
  2. Script error: No such module "citation/CS1".
  3. Script error: No such module "citation/CS1". Raw Moments at Math-world
  4. Script error: No such module "citation/CS1".
  5. Script error: No such module "Citation/CS1".
  6. Feller, W. (1957-1971). An introduction to probability theory and its applications. New York: John Wiley & Sons. 419 p.

Script error: No such module "Check for unknown parameters".

Further reading

  • Script error: No such module "citation/CS1".
  • Script error: No such module "citation/CS1".

External links

Template:Theory of probability distributions Script error: No such module "Navbox".