You must be logged in to see worked solutions.

Already have an account? Log in here.

A Markov chain is a mathematical system that evolves with time. It consists of a set of states and a matrix that contains the probabilities of transitioning from one state to another. In each time unit, the chain transitions randomly from one state to another, making the state the chain is in at any time at random variable.

For example, a person walking along the integers who moves to the left or right with equal probability is a Markov chain. The state space is the integers, and the transition probabilities are \(0.5\) for adjacent states, and 0 for all others.

You must be logged in to see worked solutions.

Already have an account? Log in here.

You must be logged in to see worked solutions.

Already have an account? Log in here.

Any general stochastic process can be made to satisfy the Markov property by altering the state space (and adding probabilities for any new states). In this way, it can "turn into" a Markov chain. Given a stochastic process with state space \(S\), which of the following methods would create a Markov chain that models the same process?

- Alter \(S\) to instead contain sequences of states, the \(k^\text{th}\) element of which tells where the Markov chain was at step \(k\).
- Change \(S\) to contain pairs of states \((i, \, j)\) instead. That way, the Markov chain can know both the present state and the previous state.
- Modify \(S\) to contain only one state, but make its transition probabilities contain all the information about the stochastic process.
- Don't change \(S\), but make the transition probabilities satisfy the Markov property.
- Add a large number of "in-between" states to \(S\), to create a bunch of little steps satisfying the Markov property.

You must be logged in to see worked solutions.

Already have an account? Log in here.

The Golden State Warriors are strongly affected by morale, and they win games according to the probabilities in the following table:

States | Win Tomorrow | Lose Tomorrow |

Won Today | \(\tfrac{4}{5}\) | \(\tfrac{1}{5}\) |

Lost Today | \(\tfrac{1}{3}\) | \(\tfrac{2}{3}\) |

If they lose the first two games in a "best of \(7\)" (first to \(4\)) series, what is the probability that they will be able to come back and win it?

You must be logged in to see worked solutions.

Already have an account? Log in here.

In the Markov chain pictured, what is the probability that a process beginning at state A will be back at state A after \(3\) moves?

You must be logged in to see worked solutions.

Already have an account? Log in here.

×

Problem Loading...

Note Loading...

Set Loading...