Documente Academic
Documente Profesional
Documente Cultură
Farzin Zareian
University of California-Irvine
1 / 20
2 / 20
Perhaps the most common measure used to describe the central location
of a random variable, X , is its mean, which is usually denoted X or E [X ].
The mean of a random variable is defined as the first moment of its PMF
or PDF, i.e.,
X
For discrete random variables: X =
xpX (x)
allx
xfX (x)
3 / 20
4 / 20
5 / 20
Expectation
Z Z
For vector of random variables: E [g (X)] =
...
g (x)fX (x)dx
|
{z
}
n-fold
6 / 20
Expectation
If X = [X1 , X2 ]
i=1
T
Z
then: E [g (X1 )] =
...
| {z
k-fold
7 / 20
8 / 20
Variance
Consequently
First moment, m = 1, E (X X )1 = 0
Second moment, m = 2, E (X X )2 = Var(X ) = X2
9 / 20
Variance
10 / 20
Coefficient of skewness
Because the third central moment, E (X X )3 , retains the sign of the
deviation from the mean, it provides useful information about the
non-symmetry of a probability distribution about its mean. Coefficient of
skewness, is used to quantify this property. For symmetric distributions,
X = 0. When X > 0, the probability distribution has a longer tail to the
right and is described as right-skewed. Conversely, X < 0 implies a longer
left tail, i.e., left-skewed.
E (X X )3
X =
|X3 |
11 / 20
Kurtosis coefficient
The fourth central moment can be used to define the Kurtosis coefficient
which quantifies the flatness of the probability distribution.
E (X X )4
X =
|X4 |
It should be noted that the above moments only provide a partial
description of a probability distribution. It can be shown that all moments
must be known in order to have complete knowledge about a random
variable.
12 / 20
|
{z
}
n-fold
...
(x1 1 )m1 (x2 2 )m2 ...(xn n )mn fX (x)dx
| {z }
n-fold
13 / 20
E [Xi Xj ] =
14 / 20
|Cov (Xi , Xj )|
p
Var (Xi )Var (Xj )
defined as: ij =
= i ij j
i j
1 ij 1
Independent random variables: E [Xi Xj ] = E [Xi ]E [Xj ] and ij = 0.
Independent random variables are uncorrelated. But the reverse is not
necessarily true.
15 / 20
Example 18
1
ba
16 / 20
When dealing with more than two random variables, it is useful, if not
necessary, to introduce a matrix notation for the first and second
moments. For an n-vector of random variables, X = [X1 , X2 , ..., Xn ]T , we
define the n 1 mean vector:
MX = [1 , 2 , ..., n ]T
and the n n covariance matrix,
Var (X1 )
Cov (X1 , X2 ) . . . Cov (X1 , Xn )
..
Cov (X1 , X2 )
Var (X2 )
.
XX =
..
.
.
.
.
.
.
.
Cov (Xn , X1 )
...
...
Var (Xn )
17 / 20
XX
Var (X1 )
Cov (X1 , X2 ) . . . Cov (X1 , Xn )
..
Cov (X1 , X2 )
Var (X2 )
.
=
..
..
.
..
.
.
Cov (Xn , X1 )
...
...
Var (Xn )
XX
12
12 1 2 . . . 1n 1 n
..
12 1 2
22
.
=
..
.
.
..
..
.
n1 n 1
...
...
n2
18 / 20
DX =
.
.
.
n
1 12 . . . 1n
..
21 1
.
RXX = .
.. = Correlation Coefficient Matrix
..
..
.
.
n1 . . . . . . 1
It can be shown that both of these matrices are, in general, positive
definite. The only exception to this rule occurs when a linear
dependence exists between any subset of random variables, in which
case the covariance matrix and the correlation coefficient matrix are
singular.
19 / 20
20 / 20