Markov chain

A **Markov chain** (named for Andrey Markov) is a sequence of random variable taking values in the state space of the chain, with the property that the probability of moving to the next state depends only upon the current state:

$Pr(X_{n + 1} = x | X_1 = x_1, X_2 = x_2,..., X_n = x_n) = Pr(X_{n + 1} = x | X_n = x_n).$

A Markov chain can also be desribed as a coalgebra for the endofunctor on Set which maps a set $X$ to the set of probability distributions on $X$.

Ian Durham: I assume that a Markov chain can be represented as a directed graph in some way and thus can be used to generate a free category of some sort. Is this a correct assumption?

Eric: I’m pretty sure that is the case for *finite* Markov chains.

Ian Durham: Hmmm. I’ll have to think about that.

For now, see Wikipedia.

Revised on December 4, 2010 18:22:08
by Toby Bartels
(98.16.130.190)