Markov Chain
A Markov chain or Markov process is a stochastic process describing a sequence of possible events in which the probability of each event depends.
A Markov chain is a mathematical system that experiences transitions from one state to another according to certain probabilistic rules. The defining characteristic of a Markov chain is that no matter how the process arrived at its present state, the possible future states are fixed. In other words, the probability of transitioning to any particular state is dependent solely on the current state and time elapsed. The state space, or set of all possible states, can be anything: letters, numbers, weather conditions, baseball scores, or stock performances.
Markov chains may be modeled by finite state machines , and random walks provide a prolific example of their usefulness in mathematics. They arise broadly in statistical and information-theoretical contexts and are widely employed in economics , game theory , queueing (communication) theory , genetics , and finance . While it is possible to discuss Markov chains with any size of state space, the initial theory and most applications are focused on cases with a finite (or countably infinite) number of states.
Markov chains are a probabilistic technique that can be used to model systems with a finite number of states and transitions between them. They are used in many fields, including:?