[Math] Markov process vs. markov chain vs. random process vs. stochastic process vs. collection of random variables

markov chainsmarkov-processrandomstochastic-processes

I'm trying to understand each of the above terms, and I'm having a lot of trouble deciphering the difference between them.

According to Wikipeda:

  • A Markov chain is a memoryless, random process.

  • A Markov process is a stochastic process, which exhibits the Markov property.

  • The Markov property is the memorylessness of a stochastic property.

  • A stochastic process is a random process, which is a collection of random variables.

  • And finally, random variables are those determined by chance instead of other variables, which seems to mean explicitly that they are memoryless.

Thus, it seems that stochastic process, random process, Markov chain, and Markov process are all the exact same thing… which is a collection of random variables, which are memory-less, which means they exhibit the Markov property.

Best Answer

The difference between Markov chains and Markov processes is in the index set, chains have a discrete time, processes have (usually) continuous.

Random variables are much like Guinea pigs, neither a pig, nor from Guinea. Random variables are functions (which are deterministic by definition). They are defined on probability space which most often denotes all possible outcomes of your experiment/model. In schools their value set is almost always a subset of $\mathbb{R}$.

Sequences of random variables don't need to be memoryless, e.g. sequences of random variables that denote some cumulative usually aren't memoryless. On the other hand, for example, sequences of independent identically distributed random variables do not depend on time at all, and so they have to be memoryless. Those two examples are something like extremes, where the next variable in the sequence depends on all of the previous (in the former example), or none of them (in the latter). The Markov property tells us, they may depend, but if they do, it does not give us any more information (e.g. in the case of discrete time, that is, Markov chains, it means that the next can be determined only using the current and nothing else). Finally, note that there is a difference between "does not depend" and "does not give us any new information", for example consider a Markov chain defined on the set of finite binary sequences where each step adds a (uniformly) random bit. Clearly, the next state does depend on all the previous "coin flips" (these are embedded in the bits of sequence prefix), but the current state already contains everything we need.

I hope it explained something $\ddot\smile$