Documente Academic
Documente Profesional
Documente Cultură
x f
X(t)
(x) dx
m
X
(t) is a function of time. It species the average behavior (or the
trend in the behavior) of X(t) over time.
Auto-correlation function: R
X
(t
1
, t
2
) is dened as the correlation
between the two time samples X
t
1
= X(t
1
) and X
t
2
= X(t
2
)
R
X
(t
1
, t
2
) = E[X
t
1
X
t
2
]
Properties:
In general, R
X
(t
1
, t
2
) depends on both t
1
and t
2
.
For real processes, R
X
(t
1
, t
2
) is symmetric
R
X
(t
1
, t
2
) = R
X
(t
2
, t
1
)
ES150 Harvard SEAS 6
For any t, t
1
and t
2
R
X
(t, t) = E[X
2
t
] 0
|R
X
(t
1
, t
2
)|
_
E[X
2
t
1
]E[X
2
t
2
]
Processes with E[X
2
t
] < for all t is called second-order.
Auto-covariance function: C
x
(t
1
, t
2
) is dened as the covariance
between the two time samples X(t
1
) and X(t
2
)
C
X
(t
1
, t
2
) = E [{X
t
1
m
X
(t
1
)}{X
t
2
m
X
(t
2
)}]
= R
X
(t
1
, t
2
) m
X
(t
1
)m
X
(t
2
)
The variance of X(t) can be obtained as
var(X
t
) = E[{X(t) m
X
(t)}
2
] = C
X
(t, t)
var(X
t
) is a function of time and is always non-negative.
The correlation coecient function:
X
(t
1
, t
2
) =
C
X
(t
1
, t
2
)
_
C
X
(t
1
, t
1
)
_
C
X
(t
2
, t
2
)
X
(t
1
, t
2
) is a function of times t
1
and t
2
. It is also symmetric.
ES150 Harvard SEAS 7
Examples: Find the mean and autocorrelation functions of the
following processes:
a) X(t) = cos(2ft +) , U[, ]
b) X
n
= Z
1
+. . . +Z
n
, n = 1, 2, . . .
where Z
i
are i.i.d. with zero mean and variance
2
.
ES150 Harvard SEAS 8
Multiple random processes:
Cross-covariance and cross-correlation functions
For multiple random processes:
Their joint behavior is completely specied by the joint distributions
for all combinations of their time samples.
Some simpler functions can be used to partially specify the joint behavior.
Consider two random processes X(t) and Y (t).
Cross-correlation function:
R
X,Y
(t
1
, t
2
) = E[X
t
1
Y
t
2
]
If R
X,Y
(t
1
, t
2
) = 0 for all t
1
and t
2
, processes X(t) and Y (t) are
orthogonal.
Unlike the auto-correlation function, the cross-correlation function
is not necessarily symmetric.
R
X,Y
(t
1
, t
2
) = R
X,Y
(t
2
, t
1
)
ES150 Harvard SEAS 9
Cross-covariance function:
C
X,Y
(t
1
, t
2
) = E[{X
t
1
m
X
(t
1
)}{Y
t
2
m
Y
(t
2
)}]
= R
X,Y
(t
1
, t
2
) m
X
(t
1
)m
Y
(t
2
)
If C
X,Y
(t
1
, t
2
) = 0 for all t
1
and t
2
, processes X(t) and Y (t) are
uncorrelated.
Two processes X(t) and Y (t) are independent if any two vectors of
time samples, one from each process, are independent.
If X(t) and Y (t) are independent then they are uncorrelated:
C
X,Y
(t
1
, t
2
) = 0 t
1
, t
2
(the reverse is not always true).
Example: Signal plus noise
Y (t) = X(t) +N(t)
where X(t) and N(t) are independent processes.
ES150 Harvard SEAS 10
Stationary random processes
In many random processes, the statistics do not change with time. The
behavior is time-invariant, even though the process is random. These are
called stationary processes.
Strict-sense stationarity:
A process is nth order stationary if the joint distribution of any set
of n time samples is independent of the placement of the time origin.
[X(t
1
), . . . , X(t
n
)] [X(t
1
+), . . . , X(t
n
+)]
For a discrete process:
[X
1
, . . . , X
n
] [X
1+m
, . . . , X
n+m
] m
A process that is nth order stationary for every integer n > 0 is said
to be strictly stationary, or just stationary for short.
Example: The i.i.d. random process is stationary.
Strict stationarity is a strong requirement.
ES150 Harvard SEAS 11
First-order stationary processes: f
X(t)
(x) = f
X
(x) for all t. Thus
m
X
(t) = m t
var(X
t
) =
2
t
Second-order stationary processes:
f
X(t
1
),X(t
2
)
(x
1
, x
2
) = f
X(t
1
+),X(t
2
+)
(x
1
, x
2
)
The second-order joint pdf (pmf) depends only on the time
dierence t
2
t
1
. This implies
R
X
(t
1
, t
2
) = R
X
(t
2
t
1
)
C
X
(t
1
, t
2
) = C
X
(t
2
t
1
)
ES150 Harvard SEAS 12
Wide-sense stationary random processes
X(t) is wide-sense stationary (WSS) if the following two properties
both hold:
m
X
(t) = m t
R
X
(t
1
, t
2
) = R
X
(t
2
t
1
) t
1
, t
2
WSS is a much more relaxed condition than strict-sense stationarity.
All stationary random processes are WSS. A WSS process is not
always strictly stationary.
Example: Sequence of independent r.v.s
X
n
= 1 with probability
1
2
for n even
X
n
= 1/3 and 3 with probabilities
9
10
and
1
10
for n odd
Properties of a WSS process:
R
X
(0) is the average power of the process
R
X
(0) = E[X(t)
2
] 0
R
X
(0) thus is always positive.
ES150 Harvard SEAS 13
R
X
() is an even function
R
X
() = R
X
()
R
X
() is maximum at = 0
|R
X
()| R
X
(0)
If R
X
(0) = R
X
(T) then R
X
() is periodic with period T
if R
X
(0) = R
X
(T) then R
X
() = R
X
( +T)
R
X
() measures the rate of change of the process
P [|X(t +) X(t)| > ]
2 (R
X
(0) R
X
())
2
If a Gaussian process is WSS, then it is also strictly stationary.
A WWS Gaussian process is completely specied by the constant
mean m and covariance C
X
().
WSS processes play a crucial role in linear time-invariant systems.
ES150 Harvard SEAS 14
Cyclostationary random processes
Many processes involves the repetition of a procedure with period T.
A random process is cyclostationary if the joint distribution of any set
of samples is invariant over a time shift of mT (m is an integer)
[X(t
1
), . . . , X(t
n
)] [X(t
1
+mT), . . . , X(t
n
+mT)] m, n, t
1
, . . . , t
n
A process is wide-sense cyclostationary if for all integer m
m
X
( +mT) = m
X
()
R
X
(t
1
+mT, t
2
+mT) = R
X
(t
1
, t
2
)
If X(t) is WSS, then it is also wide-sense cyclostationary.
We can obtain a stationary process X
s
(t) from a cyclostationary
process X(t) as
X
s
(t) = X(t +) , U[0, T]
If X(t) is wide-sense cyclostationary then X
s
(t) is WSS.
ES150 Harvard SEAS 15
Time averages and ergodicity
Sometimes we need to estimate the parameters of a random process
through measurement.
A quantity obtainable from measurements is the ensemble average. For
example, an estimate of the mean is
m
X
(t) =
1
N
N
i=1
X(t,
i
)
where
i
is the ith outcome of the underlying random experiment.
In general, since m
X
(t) is a function of time, we need to perform N
repetitions of the experiment at each time t to estimate m
X
(t).
If the process is stationary, however, then m
X
(t) = m for all t. Then
we may ask if m can be estimated based on the realization (over time)
of a single outcome alone.
We dene the time average over an interval 2T of a single realization as
X(t)
T
=
1
2T
_
T
T
X(t, ) dt
ES150 Harvard SEAS 16
Question: When does the time average converge to the ensemble
average?
Example 1: If X
n
= X(n, ) is a stationary, i.i.d. discrete-time random
process with mean E[X
n
] = m, then by the strong LLN
1
N
N
i=1
X
i
a.s.
m as N
= Convergence.
Example 2: X(t) = A for all t, where A is a zero-mean random
variable. Then X(t) is stationary and m
X
= E[A] = 0 for all t, but
X(t)
T
=
1
2T
_
T
T
Adt = A
= No convergence.
Ergodicity let us characterize this convergence for a larger class of
random processes.
ES150 Harvard SEAS 17
Ergodicity of a WSS process
Consider a WSS random process X(t) with mean m. X(t) is
mean-ergodic if
X(t)
T
m as T
Notes:
Because of stationarity, the expected value of X(t)
T
is
E [X(t)
T
] = E
_
1
2T
_
T
T
X(t) dt
_
=
1
2T
_
T
T
E[X(t)] dt = m
Mean-ergodic denition therefore implies that X(t)
T
approaches
its mean as T .
Mean-ergodic theorem: The WSS process X(t) is mean-ergodic in the
mean-square sense, that is
lim
T
E
_
(X(t)
T
m)
2
_
= 0
if and only if its covariance satises
lim
T
1
2T
_
T
T
_
1
|u|
2T
_
C
X
(u) du = 0
ES150 Harvard SEAS 18