自相关函数是否完全描述了随机过程?


31

随机过程是否由其自相关函数完全描述?

如果不是,则需要哪些其他属性?

Answers:


44

随机过程的完整描述是什么意思?好吧,在数学上,一个随机过程是随机变量的集合,在索引集,每个瞬时都有一个,其中通常是是整条实线或正实线,完整的描述意味着对于每个整数和时刻,我们知道的(联合)分布所述随机变量,,{ X Ť } {X(t):tT}t Ť TŤTÑ 1 n1Ñ n12... ÑŤt1,t2,,tnT Ñ nX 1X(t1)X 2X(t2)... X Ñ,X(tn)。这是大量信息:我们需要知道每个时刻的的CDF,所有时刻选择的和的(二维)联合CDF。和,,和等三维)CDF 等,等等。X(t)X(t)ttX(t1)X(t1)X(t2)X(t2)t1t1t2t2X(t1)X(t1)X(t2)X(t2)X(t3)X(t3)

因此,人们自然会寻找更简单的描述和更严格的模型。当过程对于时间原点的变化不变时,就会发生一种简化。这意味着

  • 该过程中的所有随机变量具有相同的CDF:对于所有。FX(t1)(x)=FX(t2)(x)FX(t1)(x)=FX(t2)(x)t1,t2t1,t2
  • 任何两个随机变量相隔一定的时间指定量具有相同的关节CDF作为任何其他对随机变量的分离由相同量的时间。例如,随机变量和相隔秒,随机变量和也因此X(t1)X(t1)X(t1+τ)X(t1+τ)ττX(t2)X(t2)X(t2+τ)X(t2+τ)FX(t1),X(t1+τ)(x,y)=FX(t2),X(t2+τ)(x,y)FX(t1),X(t1+τ)(x,y)=FX(t2),X(t2+τ)(x,y)
  • 隔开和任意三个随机变量,,具有与,,也将和隔开,X(t1)X(t1)X(t1+τ1)X(t1+τ1)X(t1+τ1+τ2)X(t1+τ1+τ2)τ1τ1τ2τ2X(t2)X(t2)X(t2+τ1)X(t2+τ1)X(t2+τ1+τ2)X(t2+τ1+τ2)τ1τ1τ2τ2
  • 对于所有多维CDF ,等等。有关多维案例的详细信息,请参见例如Peter K.的答案。

实际上,随机过程的概率描述不依赖于我们选择在时间轴上称呼原点的方式:移动所有时刻 t 1t 2... t nt1,t2,,tn一定的固定量 ττ t 1 + τ t 2 + τ t n + τt1+τ,t2+τ,,tn+τ对随机变量给出相同的概率描述。此属性称为严格意义上的平稳性 享有此属性的随机过程称为严格平稳的随机过程,或更简单地说,称为平稳的随机过程。

请注意,严格的平稳性本身并不需要任何特定形式的CDF。例如,它没有说所有变量都是高斯变量。

形容词严格建议可以定义较宽松的平稳形式。如果Ñ Nth的阶联合CDF X 1X 2... X ÑX(t1),X(t2),,X(tN)是一样的Ñ Nth阶的关节CDF X 1 + τ X t 2 + τ X tÑ + τ X(t1+τ),X(t2+τ),,X(tN+τ)为所有的选择12... Ñt1,t2,,tN ττ,则该随机过程被认为是静止的,以顺序 ÑN和作为被称为 Ñ Nth阶平稳随机过程。请注意,一个 Ñ Nth阶平稳随机过程也静止顺序 Ñn每个正 Ñ < Ñn<N。(这是因为 Ñ nth阶关节CDF是极限 Ñ Nth阶CDF作为 Ñ- ÑNn的参数接近:的一般化 ˚F XX = LIM ý →交通 ˚F X ÿX ÿ FX(x)=limyFX,Y(x,y))。那么严格平稳的随机过程是对所有阶数 NN都静止的随机过程。

如果一个随机过程至少在11阶上是平稳的,则所有X t X(t)都具有相同的分布,因此,假设存在均值,则对于所有tE [ X t ] = μE[X(t)]=μ都是相同的。类似地, E [ X t 2 ]对于所有t都是相同的,称为过程的。所有物理过程都具有有限的能力,因此通常假定 E [tE[(X(t))2]tX t 2 ] < ∞,E[(X(t))2]<在这种情况下,尤其是在较早的工程文献中,该过程称为二阶过程。名称的选择是不幸的,因为它邀请的困惑与二阶 平稳(参见我的上stats.SE这个答案),所以在这里我们将调用一个过程,这 é [ X 牛逼2 ]E[(X(t))2]是有限的对于所有 tt(是否为 E [ X t 2 ]E[(X(t))2]是一个常数),为有限功率处理和避免这种混淆。但请再次注意

一阶平稳过程不必是有限幂过程。

考虑一个随机过程,该过程的阶数为22。现在,由于的联合分布X 1X(t1)X 1 + τ X(t1+τ)是一样的联合分布函数X 2X(t2)X 2 + τ X(t2+τ)È [ X 1X t 1 + τ ] = E [ Xt 2X t 2 + τ ]E[X(t1)X(t1+τ)]=E[X(t2)X(t2+τ)],该值仅取决于 ττ。这些期望对于有限功率过程是有限的,它们的值称为过程的自相关函数: R Xτ = E [ X t X t + τ ]RX(τ)=E[X(t)X(t+τ)] ττ的函数,时间间隔随机变量 X t X(t) X t+ τ X(t+τ),并且完全不依赖于 tt。另请注意, E [ X t X t + τ ] = E [ X t + τ X t ] = E [ X t + τ X t + τ - τ ] = R X- τ

E[X(t)X(t+τ)]=E[X(t+τ)X(t)]=E[X(t+τ)X(t+ττ)]=RX(τ),
因此自相关函数是其参数的偶函数。

有限幂二阶平稳随机过程具有以下性质

  1. 它的平均值E [ X t ]E[X(t)]是一个常数
  2. 它的自相关函数R Xτ = E [ X t X t + τ ]RX(τ)=E[X(t)X(t+τ)]ττ的函数,是随机变量X t X(t)X t + τ X(t+τ)的时间间隔,而不是完全取决于tt

平稳性的假设在某种程度上简化了随机过程的描述,但是,对于对从实验数据中构建模型感兴趣的工程师和统计学家而言,估算所有这些CDF并非易事,特别是当只有一部分样本路径时(或实现)X x(t)在其上的测量可以进行。相对容易进行的两个测量(因为工程师已经在其工作台上配备了必要的仪器(或软件库中的MATLAB / Python / Octave / C ++中的程序))的直流值 1Ť Ť 0 Xdt1TT0x(t)dt of x(t)x(t) and the autocorrelation function Rx(τ)=1TT0x(t)x(t+τ)dtRx(τ)=1TT0x(t)x(t+τ)dt (or its Fourier transform, the power spectrum of x(t)x(t)). Taking these measurements as estimates of the mean and the autocorrelation function of a finite-power process leads to a very useful model that we discuss next.


A finite-power random process is called a wide-sense-stationary (WSS) process (also weakly stationary random process which fortunately also has the same initialism WSS) if it has a constant mean and its autocorrelation function RX(t1,t2)=E[X(t1)X(t2)]RX(t1,t2)=E[X(t1)X(t2)] depends only on the time difference t1t2t1t2 (or t2t1t2t1).

Note that the definition says nothing about the CDFs of the random variables comprising the process; it is entirely a constraint on the first-order and second-order moments of the random variables. Of course, a finite-power second-order stationary (or NthNth-order stationary (for N>2N>2) or strictly stationary) random process is a WSS process, but the converse need not be true.

A WSS process need not be stationary to any order.

Consider, for example, the random process {X(t):X(t)=cos(t+Θ),<t<}{X(t):X(t)=cos(t+Θ),<t<} where ΘΘ takes on four equally likely values 0,π/2,π0,π/2,π and 3π/23π/2. (Do not be scared: the four possible sample paths of this random process are just the four signal waveforms of a QPSK signal). Note that each X(t)X(t) is a discrete random variable that, in general, takes on four equally likely values cos(t),cos(t+π/2)=sin(t),cos(t+π)=cos(t)cos(t),cos(t+π/2)=sin(t),cos(t+π)=cos(t) and cos(t+3π/2)=sin(t)cos(t+3π/2)=sin(t), It is easy to see that in general X(t)X(t) and X(s)X(s) have different distributions, and so the process is not even first-order stationary. On the other hand, E[X(t)]=14cos(t)+14(sin(t))+14(cos(t))+14sin(t)=0

E[X(t)]=14cos(t)+14(sin(t))+14(cos(t))+14sin(t)=0
for every tt while E[X(t)X(s)]=14[cos(t)cos(s)+(cos(t))(cos(s))+sin(t)sin(s)+(sin(t))(sin(s))]=12[cos(t)cos(s)+sin(t)sin(s)]=12cos(ts).
E[X(t)X(s)]=14[cos(t)cos(s)+(cos(t))(cos(s))+sin(t)sin(s)+(sin(t))(sin(s))]=12[cos(t)cos(s)+sin(t)sin(s)]=12cos(ts).
In short, the process has zero mean and its autocorrelation function depends only on the time difference tsts, and so the process is wide sense stationary. But it is not first-order stationary and so cannot be stationary to higher orders either.

Even for WSS processes that are second-order stationary (or strictly stationary) random processes, little can be said about the specific forms of the distributions of the random variables. In short,

A WSS process is not necessarily stationary (to any order), and the mean and autocorrelation function of a WSS process is not enough to give a complete statistical description of the process.

Finally, suppose that a stochastic process is assumed to be a Gaussian process ("proving" this with any reasonable degree of confidence is not a trivial task). This means that for each tt, X(t)X(t) is a Gaussian random variable and for all positive integers n2n2 and choices of nn time instants t1t1, t2t2, ,tn,tn, the NN random variables X(t1), X(t2), ,X(tn) are jointly Gaussian random variables. Now a joint Gaussian density function is completely determined by the means, variances, and covariances of the random variables, and in this case, knowing the mean function μX(t)=E[X(t)] (it need not be a constant as is required for wide-sense-stationarity) and the autocorrelation function RX(t1,t2)=E[X(t1)X(t2)] for all t1,t2 (it need not depend only on t1t2 as is required for wide-sense-stationarity) is sufficient to determine the statistics of the process completely.

If the Gaussian process is a WSS process, then it is also a strictly stationary Gaussian process. Fortunately for engineers and signal processors, many physical noise processes can be well-modeled as WSS Gaussian processes (and therefore strictly stationary processes), so that experimental observation of the autocorrelation function readily provides all the joint distributions. Furthermore since Gaussian processes retain their Gaussian character as they pass through linear systems, and the output autocorrelation function is related to th input autocorrelation function as Ry=h˜hRX

so that the output statistics can also be easily determined, WSS process in general and WSS Gaussian processes in particular are of great importance in engineering applications.


Could you, please, comment on "White Noise" in that sense? By definition the Autocorrelation at τ=0 is the variance of the random variables. Does it mean that AWGN (Additive White Gaussian Noise) has infinite variance? I ask it because usually people write n(t) N(0,N0/2), is wrong? Should it be written n(t) N(0,δ(0)N0/2)? Thanks.
Royi

1
@Drazick Please ask a separate question.
Dilip Sarwate

1
This is a fantastic mini-course in the definition of stationary processes. I've never seen anything like it--laid out so methodically and clearly. Community Wiki?
abalter

@Dilip Sarwate Excuse me for my ignorance. In the example. Why is E[X(t)]=0 for all t ? Did you assume ergodicity? How did you derive the probability density function of X(t) from the probability density function of theta to compute the expected value? E[X(t)X(s)] = E[cos(t+theta)*cos(s+theta)] right? Which steps did you take to simplify this expression and get to what you wrote? Thanks
VMMF

1
@VMMF There is NO ergodicity used. X(t)=cos(t+Θ) is a discrete random variable because Θ is a discrete random variable and it takes on values ±cos(t) and ±sin(t) with equal probability 14. Ergo, E[X(t)]=0. X(t)X(s) takes on values cos(t)cos(s), (cos(t))(cos(s))=cos(t)cos(s), sin(t)sin(s) and (sin(t))(sin(s))=sin(t)sin(s) with equal probability 14. Hence, E[X(t)(X(s)]=12(cos(t)cos(s)+sin(t)sin(s))=12cos(ts). Hence,
Dilip Sarwate
By using our site, you acknowledge that you have read and understand our Cookie Policy and Privacy Policy.
Licensed under cc by-sa 3.0 with attribution required.