Inverting the Steady State of a Markov Chain

Recently while working on a time-series rare-event prediction problem I found myself wanting to generate an infinite stream of training data to test out various prediction approaches. In order to match data from my problem domain, I wanted a model that incorporated a ‘normal’ mode of operation and rare ‘failure’ events, each failure possibly preceded by one or more ‘alarm’ events, and I wanted to be able to create many such models easily. Seemed like a Markov chain would do the trick if I could just figure out a way to go from the desired steady-state distribution to the transition matrix itself. I was actually surprised to find that this is not only possible in many cases, but that the process is relatively simple to understand and to implement with the right tools. More on that later, first lets review the basics of Markov Chains.

An example of the type of Markov chain I needed to generate.

Markov Chains

Recall that a Markov chain is “a random process that undergoes transitions from one state to another on a state space.” We can represent a Markov chain using a transition matrix, and for our purposes we will use a right-stochastic matrix (meaning that all of its entires are in [0..1] and all of its rows sum to 1.0).

An example transition matrix is given below:

$P=\left(\begin{array}{cccc}.95&.02&.02&.01\\ .6&.3&.1&0\\ .6&0&.3&.1\\.2&0&0&.8\end{array}\right)$

The matrix represents the transition probabilities between the states. If we represent our current state as a vector of probabilities, and represent being in state one by assigning probability 1.0 to the first entry in our state vector, we could compute our next state by multiplying our state vector with the transition matrix (I transposed the vectors here for display purposes only, they are row vectors):

$\left(\begin{array}{cccc}1.0\\ 0\\ 0\\ 0\end{array}\right)^T \cdot P=\left(\begin{array}{cccc}.95\\ .02\\ .02\\ .01\end{array}\right)^T$

So after one step we have a 95% chance of being in state 1, a 2% chance of being in state 2, a 2% chance of being in state 3, and a 1% chance of being in state 4. To see our chances of being in each state after $n$ iterations we can apply the same approach but instead of using $P$, we use $P^n$. By extension if we take $P^{\infty}$ we will get what is called the steady-state vector of the Markov chain. This represents the long-term probabilities of being in a particular state. It is so named because it isn’t changed by the matrix $P$. In other words $s$ is a left-eigenvector of $P$ with the eigenvalue of 1.0.

$sP=1.0*s$

So if we want to represent a rare event preceded by some pre-event states we could model it as a Markov chain using something like the below, where $s$ represents the steady state distribution of $P$.

$s=\left(\begin{array}{cccc}0.886\\ 0.025\\ 0.028\\ 0.058\end{array}\right),P=\left(\begin{array}{cccc}.95&.02&.02&.01\\ .6&.3&.1&0\\ .6&0&.3&.1\\.2&0&0&.8\end{array}\right)$

The first row in $P$ represents the ‘normal’ state. It mostly loops back into itself, occasionally transitioning to the other states. The last row is the ‘failure’ state, which mostly loops back to itself, but after a time it will go back to state one. The other two states are a “chain” that links the normal state to the failure state in one or more steps. These chains allow us to gradually move to the  failure mode, which is required to make predicting it possible.

The main goal here is to create some rare failure event data so we start by choosing $s$ to assign a reasonable amount of probability to state 1 (say 80%) and a very small probability to the failure state (maybe 1e-5). The rest of the probability mass is randomly distributed over the other states. So we have something like this (values rounded for readability):

$s=\left(\begin{array}{cccc}.8\\ .16673\\ .03326\\ .00001\end{array}\right)$

So the question is, can we build a transition matrix $P$ such that $s$ is the steady-state vector?

It turns out that this is an under-constrained problem and there may be many solutions, if one exists at all.

Rules of Thumb

1. P must be irreducible – A Markov chain is reducible if a state $s_i$ exists that is inaccessible from some other state $s_j$.  In other words it has more than one non-communicating equivalence class. A reducible Markov chain will, in general, have multiple steady state distributions. Not good for our purposes.

2. P must be strongly connected – Assuming an irreducible Markov chain, in order for a unique stationary distribution to exist we require that all states be positive recurrent. A state is positive recurrent if the expected value of the recurrence time (the number of transitions until the state is re-entered) is finite. This is equivalent to the requirement that the graph be strongly connected. The graph is strongly connected if every state is reachable from every other state.

3. No zeros in target distribution – Given all of our previous requirements, if a stationary distribution $\boldsymbol{s}$ exists for some $P$, then it must be the case that all $s_i$ are strictly positive, so we cannot allow any zeros when setting up our target distribution.

There may be other nuances regarding when this succeeds or fails, but I have found that if I follow these rules this technique works, and works pretty fast, most of the time (once the learning rate is tuned properly).

Cost Function

1. Right Stochastic Constraint – The first thing we need to do is ensure that our matrix $P$ is right-stochastic. That is, its entries are all probabilities and that each row sums to 1.0. Let $\boldsymbol{1}$ be a vector of length $n$ and let $P$ be an $n \times n$ matrix.  Then all the rows sum to 1.0 if we can satisfy:

$\begin{array}{cc}P *\boldsymbol{1} - \boldsymbol{1} = \boldsymbol{0}\end{array}$

2. All Values Non-Negative – We also need all of our entries to be probabilities. If we assume that the rows sum to one (our previous constraint) then we simply need to ensure that all the entries are non-negative. Let $I_N$ be an indicator function that is equal to 1 if its parameter is negative, and 0 otherwise. Then all our entries are non-negative if we can satisfy:

$\begin{array}{cc}\sum_{i,j} P_{ij} * -I_N(P_{ij})=0\end{array}$

3. Eigenvector – Last, but certainly not least, we want $s$ to be an eigenvector of $P$. Rearranging the eigenvector equation we see that this requirement is satisfied by the following:

$\begin{array}{cc}s(P-I)=\boldsymbol{0}\end{array}$

4. Graph Connectivity – The constraints above are good enough to solve the fully connected case, but we will often want to specify a connectivity constraint on our transition matrix. To implement this we use a boolean connectivity matrix $C$ which has a 1 for each transition that is legal and a 0 for each transition that is illegal. We invert this matrix and then apply element-wise multiplication, leaving positive penalty values for each illegal transition that has a non-zero probability in $P$.

$\begin{array}{cc}P.*(\boldsymbol{1}-C)=\boldsymbol{0}\end{array}$

Now we can define the complete problem as follows:

$cost(s,P) = \begin{array}{cc}|s(P-I)| + \sum_{i,j} P_{ij} * -I_N(P_{ij}) + |P *\boldsymbol{1} - \boldsymbol{1}| + |P.*(\boldsymbol{1}-C)|\end{array}$

$\begin{array}{cc}\underset{P}{\text{minimize}} & cost(s,P)\end{array}$

Since our cost function is differentiable, we can use simple gradient descent to search for a solution (with all the usual caveats). We will use $P$ as our parameter and update it using the gradient and a learning rate with the following update rule:

$P = P - \alpha \nabla cost(s, P)$

In the next post we will see how to implement a simple solver for this problem in Python and see how to implement graph connectivity constraints as well.

This entry was posted in Uncategorized. Bookmark the permalink.

One Response to Inverting the Steady State of a Markov Chain

1. I learned a lot about Markov chains here. I had run across them before but only tangentially and never really delved into them or did anything with them. I followed each paragraph pretty well until the section on positive penalty values. Not sure what that is about….

Richard