博客
关于我
强烈建议你试试无所不能的chatGPT,快点击我
理解距(数学)
阅读量:6835 次
发布时间:2019-06-26

本文共 7138 字,大约阅读时间需要 23 分钟。

In , a moment is, loosely speaking, a quantitative measure of the shape of a set of points. The "second moment", for example, is widely used and measures the "width" (in a particular sense) of a set of points in one dimension or in higher dimensions measures the shape of a cloud of points as it could be fit by an . Other moments describe other aspects of a  such as how the distribution is skewed from its mean, or peaked. The mathematical concept is closely related to the concept of  in , although moment in physics is . Any distribution can be characterized by a number of features (such as the mean, the variance, the skewness, etc.), and the moments of a function describe the nature of its distribution.

The 1st moment is denoted by μ1. The first moment of the distribution of the random variable X is the expectation operator, i.e., the  (if the first moment exists).

In higher orders, the central moments (moments about the mean) are more interesting than the moments about zero. The kth , of a real-valued random variable probability distribution X, with the  μ is:

\mu_k=E((X-\mu)^k).\,

The first central moment is thus 0. The zero-th central moment, μ0 is one. See also .

Significance of the moments 距的重要性

The nth moment of a real-valued continuous function f(x) of a real variable about a value c is

\mu'_n=\int_{-\infty}^\infty (x - c)^n\,f(x)\,dx.\,\!

It is possible to define moments for  in a more general fashion than moments for real values—see . The moment of a function, without further explanation, usually refers to the above expression with c = 0.

Usually, except in the special context of the , the function f(x) will be a . The nth moment about zero of a probability density function f(x) is the  of Xn and is called a raw moment or crude moment. The moments about its mean μ are called ; these describe the shape of the function, independently of .

If f is a , then the value of the integral above is called the nth moment of the . More generally, if F is a  of any probability distribution, which may not have a density function, then the nth moment of the probability distribution is given by the 

\mu'_n = \operatorname{E}(X^n)=\int_{-\infty}^\infty x^n\,dF(x)\,

where X is a  that has this distribution and E the  or mean.

When

\operatorname{E}(|X^n|) = \int_{-\infty}^\infty |x^n|\,dF(x) = \infty,\,

then the moment is said not to exist. If the nth moment about any point exists, so does (n − 1)th moment, and all lower-order moments, about every point.

Variance方差

Main article: 

The second  about the mean is the , the positive square root of which is the  σ.

Normalized moments 正规距

The normalized nth central moment or  is the nth central moment divided by σn; the normalized nth central moment of x = E((x − μ)n)/σn. These normalized central moments are , which represent the distribution independently of any linear change of scale.

Skewness 偏态

Main article: 

The third central moment is a measure of the lopsidedness of the distribution; any symmetric distribution will have a third central moment, if defined, of zero. The normalized third central moment is called the , often γ. A distribution that is skewed to the left (the tail of the distribution is heavier on the left) will have a negative skewness. A distribution that is skewed to the right (the tail of the distribution is heavier on the right), will have a positive skewness.

For distributions that are not too different from the , the  will be somewhere near μ − γσ/6; the  about μ − γσ/2.

Kurtosis 峰态

Main article: 

The fourth central moment is a measure of whether the distribution is tall and skinny or short and squat, compared to the normal distribution of the same variance. Since it is the expectation of a fourth power, the fourth central moment, where defined, is always non-negative; and except for a , it is always strictly positive. The fourth central moment of a normal distribution is 3σ4.

The  κ is defined to be the normalized fourth central moment minus 3. (Equivalently, as in the next section, it is the fourth  divided by the square of the variance.) Some authorities do not subtract three, but it is usually more convenient to have the normal distribution at the origin of coordinates. If a distribution has a peak at the mean and long tails, the fourth moment will be high and the kurtosis positive (leptokurtic); and conversely; thus, bounded distributions tend to have low kurtosis (platykurtic).

The kurtosis can be positive without limit, but κ must be greater than or equal to γ2 − 2; equality only holds for . For unbounded skew distributions not too far from normal, κ tends to be somewhere in the area of γ2 and 2γ2.

The inequality can be proven by considering

\operatorname{E} ((T^2 -  aT)^2)\,

where T = (X − μ)/σ. This is the expectation of a square, so it is non-negative whatever a is; on the other hand, it's also a  in a. Its  must be non-positive, which gives the required relationship.

Mixed moments 混合距

Mixed moments are moments involving multiple variables.

Some examples are ,  and . While there is a unique covariance, there are multiple co-skewnesses and co-kurtoses.

Higher moments 高阶距

High-order moments are moments beyond 4th-order moments. The higher the moment, the harder it is to estimate, in the sense that larger samples are required in order to obtain estimates of similar quality.[]

Cumulants 累积距

Main article: 

The first moment and the second and third unnormalized central moments are additive in the sense that if X and Y are  random variables then

\mu_1(X+Y)=\mu_1(X)+\mu_1(Y)\,

and

\operatorname{Var}(X+Y)=\operatorname{Var}(X) + \operatorname{Var}(Y)

and

\mu_3(X+Y)=\mu_3(X)+\mu_3(Y).\,

(These can also hold for variables that satisfy weaker conditions than independence. The first always holds; if the second holds, the variables are called ).

In fact, these are the first three cumulants and all cumulants share this additivity property.

Sample moments 样本距

The moments of a population can be estimated using the sample k-th moment

\frac{1}{n}\sum_{i = 1}^{n} X^k_i\,\!

applied to a sample X1,X2,..., Xn drawn from the population.

It can be shown that the expected value of the sample moment is equal to the k-th moment of the population, if that moment exists, for any sample size n. It is thus an unbiased estimator.

Problem of moments 距的问题

The  seeks characterizations of sequences { μn : n = 1, 2, 3, ... } that are sequences of moments of some function f.

Partial moments 部分距/一边距

Partial moments are sometimes referred to as "one-sided moments." The nth order lower and upper partial moments with respect to a reference point r may be expressed as

\mu_n^-(r)=\int_{-\infty}^r (r - x)^n\,f(x)\,dx,
\mu_n^+(r)=\int_r^\infty (x - r)^n\,f(x)\,dx.

Partial moments are normalized by being raised to the power 1/n. The  may be expressed as a ratio of a first-order upper partial moment to a normalized second-order lower partial moment.

Moments in metric spaces 矩阵空间中的距

Let (Md) be a , and let B(M) be the  on M, the  generated by the d- of M. (For technical reasons, it is also convenient to assume that M is a  with respect to the  d.) Let 1 ≤ p ≤ +∞.

The pth moment of a measure μ on the  (M, B(M)) about a given point x0 in M is defined to be

\int_{M} d(x, x_{0})^{p} \, \mathrm{d} \mu (x).

μ is said to have finite pth moment if the pth moment of μ about x0 is finite for some x0 ∈ M.

This terminology for measures carries over to random variables in the usual way: if (Ω, Σ, P) is a  and X : Ω → M is a random variable, then the pth moment of X about x0 ∈ M is defined to be

\int_{M} d (x, x_{0})^{p} \, \mathrm{d} \left( X_{*} (\mathbf{P}) \right) (x) \equiv \int_{\Omega} d (X(\omega), x_{0})^{p} \, \mathrm{d} \mathbf{P} (\omega),

and X has finite pth moment if the pth moment of X about x0 is finite for some x0 ∈ M.

 

 

 

 

http://en.wikipedia.org/wiki/Moment_(mathematics)

转载于:https://www.cnblogs.com/pengkunfan/p/3998948.html

你可能感兴趣的文章
原生JS实现音乐播放器!
查看>>
hive-安装MySQL(centos6.4)
查看>>
UVa 12100 Printer Queue (习题 5-7)
查看>>
windows下安装apache zookeeper
查看>>
第三周作业
查看>>
git pull --rebase
查看>>
linux下mysql的root密码忘记解决方
查看>>
FastDFS的配置、部署与API使用解读(6)FastDFS配置详解之Storage配置(转)
查看>>
rabbitmq的web管理界面无法使用guest用户登录
查看>>
[LeetCode]题解(python):048-Rotate Image
查看>>
[LeetCode]题解(python):075-Sort Colors
查看>>
protobuf 中的嵌套消息的使用 主要对set_allocated_和mutable_的使用
查看>>
0-1背包问题
查看>>
系统的Drawable(二)-Selector
查看>>
CAS 界面根据不同的域名显示不同的界面
查看>>
java获取在各种编码下中文及英文的字符个数
查看>>
数组知识0913
查看>>
ECharts+百度地图网络拓扑应用
查看>>
LCA
查看>>
用 scp 命令通过 SSH 互传文件
查看>>