Random Vector

Definition

Column vector whose components are random variables .

Link to original

Stochastic Process

Definition

where each is a Random Variable.

Stochastic process is a collection of random variables defined on a common probability space , and whose index represents time.

Notation

Stochastic Process whose Initial Distribution is a Delta Distribution

If a random variable has a positive probability only at and otherwise , then the distribution is expressed as . Also, the probability and expectation of are expressed as , and respectively.

The probability of given some initial distribution is expressed as following

  • where is the initial state
  • where is the initial distribution

where is a HMC defined on the state space

Averages

For a sequence of random variables

Ensemble Average

Time Average

Facts

If the random variables of a stochastic process satisfy i.i.d condition, then the ensemble average is equal to time average

Summary

The Distributions of Stochastic Process

Empirical distributionStationary DistributionLimiting Distribution
Used forEvery stochastic processEvery stochastic processMarkov chain
Expression
Consideration of use single use every don’t consider
Statistical distribution?OOX
Analytic value?XOO
Related to data?OXX
Related to limit?OXO
Related to LLN?OOX
DescriptionAveraged calculated by the dataTheoretical expectationTheoretical convergence value

: the empirical distribution of with consideration of single (use time average)
: the Stationary Distribution of with consideration of every : the Limiting Distribution of obtained by the limit of without consideration of . Used with Markov Chain

where is a sequence of random variables

Link to original

Markov Property

Definition

Markov property refers to the memoryless property of a Stochastic Process.

Consider a Probability Space with a Filtration , and a Measurable Space . A Stochastic Process is said to possess the Markov property if Here, in the LHS represents all information up to time , whereas conditions only on the single value at time .

In the case where is a discrete set and , this can be reformulated as follows

Link to original

Markov Chain

Definition

A Markov chain is a Stochastic Process that satisfies Markov Property. It consists of a set of states and a Transition Probability Matrix

Summary

Conditions and Properties of Markov Chain (Finite States)

ExampleFiniteIRRAperiodicity^[a unique Limiting Distribution]^[a unique Stationary Distribution]^[Ergodic Theorem]^[Ergodicity]
OXXXXXX
Identity/Block matrixOXOXXXX
Exchange matrixOOXXOOX
Ergodic Markov ChainOOOOOOO
flowchart LR
  mc[HMC] --> finite
  
  subgraph finiteness
    finite([finite])
  end
  
  finite --> finite_mc
  finite_mc --> irreducible
  finite_mc --> reducible
  
  subgraph Irreducibility
    irreducible([irreducible])
    reducible([reducible])
  end
  irreducible --> irr_mc[Irreducible MC]
  reducible -->|divide|irr_mc

  irr_mc ---|holds|egt([Ergodic Theorem])
  irr_mc --> aperiodic
  
  subgraph Aperiodicity
    aperiodic([aperiodic])
  end
  
  aperiodic --> ergmc[Ergodic MC]

Conditions and Properties of Markov Chain (Infinite States)

IRRNatureAperiodicity^[a unique Limiting Distribution]^[a unique Stationary Distribution]^[Ergodic Theorem]^[Ergodicity]
OTransientcan’t defineXXXX
ONull recurrentno matterXXXX
OPositive recurrentXXOOX
OPositive recurrentOOOOO
flowchart LR
  mc[HMC]
  mc --> irreducible
  mc --> reducible
  
  subgraph Irreducibility
    irreducible([irreducible])
    reducible([reducible])
  end
  
  irr_mc[Irreducible MC]
  irreducible --> irr_mc
  reducible -->|divide|irr_mc
  
  irr_mc --> pr
  irr_mc --> nr
  irr_mc --> tr
  
  subgraph Recurrence
    pr([positive recurrent])
    nr([null recurrent])
    tr([transient])
  end
    
  pr_mc[Recurrent MC]
  pr --> pr_mc
  pr_mc --> aperiodic

  egt([Ergodic Theorem])
  pr_mc ---|holds|egt

  subgraph Aperiodicity
    aperiodic([aperiodic])
  end
  
  aperiodic --> ergmc[Ergodic MC]
Link to original

Homogeneous Markov Chain

Definiotion

Markov Chain that are the same for any ,

Link to original

Transition

Definition

where is the distribution of states and is Transition Matrix

Link to original

Transition Probability

Definition

where is HMC that has countable state space

The conditional probability of the Transition from state to

Facts

Link to original

Transition Probability Matrix

Definition

A transition probability matrix is a matrix whose -th element is . where each is a Transition Probability of HMC having countable state space

Facts

Since the number of elements of state space can be infinity, strictly speaking is not a matrix. However, several operations of matrix are satisfied for the

The sum of each row of a transition matrix is one

Link to original

Invariant Measure

Definition

The Measure that satisfies the above

where is a Transition Matrix of the Markov Chain

Link to original

Stationary Distribution

Definition

Invariant Measure satisfying definition of distribution

distribution that satisfies where is a Distribution defined on Measurable Space and

distribution that satisfies where is a Transition Matrix of the Markov Chain

Facts

When is a stationary distribution of the finite stochastic process , it becomes a PMF of the random variable

For the finite states case, the stationary distribution can be obtained using Eigendecomposition. The stationary distribution is the unit eigenvector^[whose Norm is one] corresponding to the eigenvalue with value one.

Finite HMC has at least one Stationary Distribution

If we set the initial distribution of the stochastic process to a stationary distribution , the stochastic process has the same distributions for every (not independent)

is a IRR-HMC PR

If is a IRR HMC, then the has a Stationary Distribution has a unique Stationary Distribution Positive recurrent

Link to original

Accessibility

Definition

, is accessible from

where is a HMC defined on the state space and the is a transition matrix of and is a -th element of

Link to original

Equivalence Relation

Definition

A binary Relation on set that satisfies the following conditions for all

Facts

The equivalence relation may be constructed from the Partition

Link to original

Communication

Definition

: are accessible from each other

where is a state space

Facts

Communication is an Equivalence Relation

Link to original

Communication Class

Definition

where The Equivalence Class of the state space by the Communication

Example

Divide Transition Matrix by Communication

where are Recurrent block matrix and is Transient block matrix

Facts

are pairwise coprime

are IRR

is PR, NR or TR

Facts

The quotient set form a Partition

Link to original
Therefore the set of communication classes^[Quotient Set] form a Partition.

Link to original

Irreducible Markov Chain

Definition

State space which has only one Communication Class. where is a HMC defined on the state space and the is a Transition Matrix of

where and is the probability of departing from and reaching after

Facts

If HMC has finite state space and is irreducible, then the unique Stationary Distribution of exists, and the probability of every state is positive.

Since Communication is a Equivalence Class, the reducible Markov chain can be partitioned into the irreducible Markov chains.

Link to original

Recurrence

Definition

Continuous Version

Let be a HMC defined on the state space , and be the return time for the state

Recurrent

Positive recurrent

Null recurrent

Transient

Incomplete Discrete Version

Let be a HMC defined on the state space , , and be the element of the transition matrix of

Recurrent

Transient

Notation

The first time to return to state

The probability of returning to from within a finite time.

Example

Let the be a HMC and be a transition matrix.

is an Irreducible Markov Chain. Whether recurrent or transient is determined by

  • : transient
  • with : positive recurrent
  • with : null recurrent

Facts

If a HMC is Irreducible Markov Chain, then every state has the same nature. In other words, it has one among the following

  • every state is transient
  • every state is null recurrent
  • every state is positive recurrent

The number of times that returning to oneself for a time approximates the inverse of .

Link to original

Period of Markov Chain

Definition

Let be a HMC defined on the state space , and the set be the set of the time steps returning from the to oneself .

The period of is the G.C.D of every element of

Link to original

Aperiodic Markov Chain

Definition

Markov Chain whose period is one.

Facts

If there is a probability of returning to oneself after 1 period or there exist non-zero diagonal elements of the Transition Matrix, then the Markov always aperiodic

Every state of finite IRR HMC has the same period.

If HMC is IRR and recurrent, then every state has the same period.

Examples

Periodic Markov Chain with 3-period

flowchart LR  
  0 -->|1| 1  
  1 -->|1| 3  
  2 -->|1| 1  
  3 -->|1/3| 0   
  3 -->|2/3| 2  

Aperiodic Markov Chain with 1-period

flowchart LR
  0 -->|1| 1
  1 -->|1/2| 0
  1 -->|1/2| 2
  2 -->|1| 3
  3 -->|1| 1
Link to original

Ergodic Theorem

Definition

A finite IRR HMC holds the following for an arbitrary function where is the unique Stationary Distribution, and is a Transition Matrix of

IRR PR HMC holds the following for a function that satisfies where is the unique Stationary Distribution, and is a Transition Matrix of

Facts

Approximate Stationary Distribution Using Ergodic Theorem (Finite)

If the Markov Chain has a unique Stationary Distribution, it can be approximated using the time average by the ergodic theorem.

Let function be , The stationary distribution can be approximated by using the ergodic theorem

Approximate Stationary Distribution Using Ergodic Theorem (Infinite)

IRR PR HMC holds

where is the time average, and is the Stationary Distribution, of

Examples

Identity Matrix

Since the IRR condition is not satisfied, there is no unique Stationary Distribution. So the Ergodic theorem cannot be used.

Exchange Matrix

Link to original

Ergodic Markov Chain

Definition

The HMC that satisfies the following conditions is an ergodic Markov chain

The HMC that satisfies the following conditions is an ergodic Markov chain

Link to original

Limiting Distribution

Definition

If exists and the every row of are the same, an arbitrary row vector of the is called the limiting distribution of the stochastic process .

where the is the transition matrix of the finite HMC

Facts

If a limiting distribution exists, it is unique.

If a finite HMC has a limiting distribution , then the limiting distribution satisfies the definition of Stationary Distribution

Ergodic Markov Chain

Ergodic Markov Chain has a unique limiting distribution

IRR TR HMC doesn’t have a limiting distribution because the converged Transition Matrix

IRR NR HMC doesn’t have a limiting distribution because the converged Transition Matrix

Link to original

Detailed Balance Condition

Definition

where is IRR HMC, is the Transition Matrix of the , and is a Distribution

The probability of the transition from the state to is equal to the probability from the to

Facts

The detailed balanced condition implies the stationarity of the distribution In other words, the distribution satisfying the detailed balanced condition is a Stationary Distribution.

Link to original

Reversible Markov Chain

Definition

The Markov Chain satisfying Detailed Balance Condition

Link to original