Moment problem

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

File:Standard deviation diagram.svg
Example: Given the mean and variance σ2 (as well as all further cumulants equal 0) the normal distribution is the distribution solving the moment problem.

In mathematics, a moment problem arises as the result of trying to invert the mapping that takes a measure μ to the sequence of moments

mn=xndμ(x).

More generally, one may consider

mn=Mn(x)dμ(x).

for an arbitrary sequence of functions Mn.

Introduction

[edit | edit source]

In the classical setting, μ is a measure on the real line, and M is the sequence {xn:n=1,2,}. In this form the question appears in probability theory, asking whether there is a probability measure having specified mean, variance and so on, and whether it is unique.

There are three named classical moment problems: the Hamburger moment problem in which the support of μ is allowed to be the whole real line; the Stieltjes moment problem, for [0,); and the Hausdorff moment problem for a bounded interval, which without loss of generality may be taken as [0,1].

The moment problem also extends to complex analysis as the trigonometric moment problem in which the Hankel matrices are replaced by Toeplitz matrices and the support of μ is the complex unit circle instead of the real line.[1]

Existence

[edit | edit source]

A sequence of numbers mn is the sequence of moments of a measure μ if and only if a certain positivity condition is fulfilled; namely, the Hankel matrices Hn,

(Hn)ij=mi+j,

should be positive semi-definite. This is because a positive-semidefinite Hankel matrix corresponds to a linear functional Λ such that Λ(xn)=mn and Λ(f2)0 (non-negative for sum of squares of polynomials). Assume Λ can be extended to [x]*. In the univariate case, a non-negative polynomial can always be written as a sum of squares. So the linear functional Λ is positive for all the non-negative polynomials in the univariate case. By Haviland's theorem, the linear functional has a measure form, that is Λ(xn)=xndμ. A condition of similar form is necessary and sufficient for the existence of a measure μ supported on a given interval [a,b].

One way to prove these results is to consider the linear functional φ that sends a polynomial

P(x)=kakxk

to

kakmk.

If mk are the moments of some measure μ supported on [a,b], then evidently

Vice versa, if (1) holds, one can apply the M. Riesz extension theorem and extend φ to a functional on the space of continuous functions with compact support Cc([a,b])), so that

By the Riesz representation theorem, (2) holds iff there exists a measure μ supported on [a,b], such that

φ(f)=fdμ

for every fCc([a,b]).

Thus the existence of the measure μ is equivalent to (1). Using a representation theorem for positive polynomials on [a,b], one can reformulate (1) as a condition on Hankel matrices.[2][3]

Uniqueness (or determinacy)

[edit | edit source]

The uniqueness of μ in the Hausdorff moment problem follows from the Weierstrass approximation theorem, which states that polynomials are dense under the uniform norm in the space of continuous functions on [0,1]. For the problem on an infinite interval, uniqueness is a more delicate question.[4] There are distributions, such as log-normal distributions, which have finite moments for all the positive integers but where other distributions have the same moments.

Formal solution

[edit | edit source]

When the solution exists, it can be formally written using derivatives of the Dirac delta function as

dμ(x)=ρ(x)dx,ρ(x)=n=0(1)nn!δ(n)(x)mn.

The expression can be derived by taking the inverse Fourier transform of its characteristic function.

Variations

[edit | edit source]

An important variation is the truncated moment problem, which studies the properties of measures with fixed first k moments (for a finite k). Results on the truncated moment problem have numerous applications to extremal problems, optimisation and limit theorems in probability theory.[3]

Probability

[edit | edit source]

The moment problem has applications to probability theory. The following is commonly used:[5]

Theorem (Fréchet-Shohat)If μ is a determinate measure (i.e. its moments determine it uniquely), and the measures μn are such that k0limnmk[μn]=mk[μ], then μnμ in distribution.

By checking Carleman's condition, we know that the standard normal distribution is a determinate measure, thus we have the following form of the central limit theorem:

CorollaryIf a sequence of probability distributions νn satisfy m2k[νn](2k)!2kk!;m2k+1[νn]0 then νn converges to N(0,1) in distribution.

See also

[edit | edit source]

Notes

[edit | edit source]
  1. ^ Schmüdgen 2017, p. 257.
  2. ^ Shohat & Tamarkin 1943.
  3. ^ a b Kreĭn & Nudel′man 1977.
  4. ^ Akhiezer 1965.
  5. ^ Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).

References

[edit | edit source]
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value). (translated from the Russian by N. Kemmer)
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).
  • Lua error in Module:Citation/CS1/Configuration at line 2172: attempt to index field '?' (a nil value).