Documente Academic
Documente Profesional
Documente Cultură
2
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Description
Sometimes we are interested in how a random
variable changes over time.
The study of how a random variable evolves
over time includes stochastic processes.
An explanation of stochastic processes in
particular, a type of stochastic process known
as a Markov chain is included.
We begin by defining the concept of a
stochastic process.
A continuous time Markov Chain (CTMC)
A Discrete time Markov Chain (DTMC)
What is a Markov Chain?
One special type of discrete-time is called a Markov
Chain.
Definition: A discrete-time stochastic process is a
Markov chain if, for t = 0,1,2 and all states
P(Xt+1 = it+1|Xt = it, Xt-1=it-1,,X1=i1, X0=i0)
=P(Xt+1=it+1|Xt = it)
Essentially this says that the probability distribution of
the state at time t+1 depends on the state at time t(it)
and does not depend on the states the chain passed
through on the way to it at time t.
In our study of Markov chains, we make further
assumption that for all states i and j and all t,
P(Xt+1 = j|Xt = i) is independent of t.
This assumption allows us to write P(Xt+1 = j|Xt
= i) = pij where pij is the probability that given
the system is in state i at time t, it will be in a
state j at time t+1.
If the system moves from state i during one
period to state j during the next period, we call
that a transition from i to j has occurred.
We call the vector q= [q1, q2,qs] the initial probability
distribution for the Markov chain.
In most applications, the transition probabilities are displayed
as an s x s transition probability matrix P. The transition
probability matrix P may be written as
p
j 1
ij 1
ashid A. Saeed
ELEXM 621 Information Theory and
8
Coding
Continuous-time Markov chains (homogeneous
case)
ashid A. Saeed
ELEXM 621 Information Theory and
9
Coding
The Gamblers Ruin Problem
At time 0, I have $2. At times 1, 2, , I play a game
in which I bet $1, with probabilities p, I win the
game, and with probability 1 p, I lose the game.
My goal is to increase my capital to $4, and as soon
as I do, the game is over. The game is also over if my
capital is reduced to 0.
Let Xt represent my capital position after the time
t game (if any) is played
X0, X1, X2, . May be viewed as a discrete-time
stochastic process
The Gamblers Ruin Problem
$0 $1 $2 $3 $4
1 0 0 0 0
1 p 0 p 0 0
P = 0 1 p 0 p 0
0 0 1 p 0 p
0 0 0 0 1
Stationary solution (steady state)
ashid A. Saeed
ELEXM 621 Information Theory and
12
Coding
The Cola Example
Suppose the entire cola industry produces only two colas.
Given that a person last purchased cola 1, there is a 90%
chance that their next purchase will be cola 1.
Given that a person last purchased cola 2, there is an 80%
chance that their next purchase will be cola 2.
1. If a person is currently a cola 2 purchaser, what is the
probability that they will purchase cola 1 two purchases from
now?
2. If a person is currently a cola 1 a purchaser, what is the
probability that they will purchase cola 1 three purchases
from now?
20%
2 1 90%
80%
The Cola Example
90%
Cola1 Cola 2
Cola 1 .90 .10
P
Cola 2 .20 .80
q1 = (.48,.52)
To illustrate the behavior of the n-step transition
probabilities for large values of n, we have
computed several of the n-step transition
probabilities for the Cola example.
This means that for large n, no matter what the
initial state, there is a .67 chance that a person
will be a cola 1 purchaser.
5.4 Classification of States in a
Markov Chain
To understand the n-step transition in more detail, we need
to study how mathematicians classify the states of a Markov
chain.
The following transition matrix illustrates most of the
following definitions. A graphical representation is shown in
the book (State-Transition diagram)
.4 .6 0 0 0
.5 .5 0 0 0
P 0 0 .3 .7 0
0 0 .5 .4 .1
0 0 0 .8 .2
Information
Processes
Memory Memoryless
Markov Processes
Human
Model
algorithm
Extensions of
Processes
20
Dr. Rashid A. Saeed
Memoryless Information Processes
For a memoryless information
source, there are no correlations
between the outputs of the source
at different times.
If all the probability distributions
are the same, the source is said to be
stationary.
21
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
22
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Memoryless information Processes are relatively simple.
More realistic information Processes have memory,
which is the property that the emission of a symbol at any instant
depends on one or more of the symbols that were generated before it.
In this section, we defined it in terms of finite state Markov
chains.
We use the state of a Markov chain to represent the
memory of the source, and
use the transitions between states to represent the next
symbol out of the source.
23
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Markov Processes
A Markov information source, is an information
source whose underlying dynamics are given by a
stationary finite Markov chain.
Markov Processes are commonly used in
communication theory, as a model of a
transmitter.
Markov Processes also occur in natural language
processing, where they are used to represent
hidden meaning in a text.
24
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Markov Processes
25
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Diagrammatic representation of a
Markov source
26
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
27
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Markov chain in Network queuing
Markov Chain for M/M/1 system
28
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Birth-death chain
29
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Markov model of a scalar passage of
music.
30
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
n-gram model,
31
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
32
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Find mth-order Markov model?
35
ELEXM 621 Information Theory and Coding
n-gram Applications
design kernels that allow machine learning
algorithms
find likely candidates for the correct spelling
of a misspelled word
improve compression in compression
algorithms
assess the probability of a given word
sequence appearing in text of a language of
interest in pattern recognition systems
36
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
37
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
38
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Stationary Distributions
39
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
The random process will not change its statistical properties with time
40
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Find the stationary distribution matrix
41
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
42
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
43
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
44
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Once the source enters either the first state or
the fourth state, it cannot leave it.
Memory Memoryless
Markov
Processes Human
Model algorithm
Extensions of
Processes
46
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
The Adjoint Source of a Markov
Source
47
memory memoryless
48
Dr. Rashid A. Saeed
49
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Proof
50
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Information
Processes
Memory Memoryless
Markov
Processes Human
Model algorithm
Extensions of
Processes
51
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Extensions of Processes
52
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
53
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
54
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
55
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Extensions of Markov Processes
56
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Extensions of Markov
M=1,Processes
n=2, p=1
57
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Design the 2nd extension?
Extensions of Markov
M=2,Processes
n=2, p=1
58
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
59
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
m=2, n=2, p=1
60
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
m=4, n=2, p=2
61
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
m=4, n=2, p=2
62
Infinite Sample Spaces
63
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Continuous Sample Space
Differential entropy or
continuous entropy
64
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
One must take care in trying to apply
properties of discrete entropy to differential
entropy,
Since probability density functions can be
greater than 1.
For example, Uniform(0,0.5) has negative
differential entropy
65
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
66
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
67
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Exercises
68
ELEXM 621 Information Theory and Coding Dr. Rashid A. Saeed
Thank You