enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Examples of Markov chains - Wikipedia

    en.wikipedia.org/wiki/Examples_of_Markov_chains

    A game of snakes and ladders or any other game whose moves are determined entirely by dice is a Markov chain, indeed, an absorbing Markov chain. This is in contrast to card games such as blackjack, where the cards represent a 'memory' of the past moves. To see the difference, consider the probability for a certain event in the game.

  3. Markov chain - Wikipedia

    en.wikipedia.org/wiki/Markov_chain

    If the Markov chain is time-homogeneous, then the transition matrix P is the same after each step, so the k-step transition probability can be computed as the k-th power of the transition matrix, P k. If the Markov chain is irreducible and aperiodic, then there is a unique stationary distribution π. [41]

  4. Transition-rate matrix - Wikipedia

    en.wikipedia.org/wiki/Transition-rate_matrix

    In probability theory, a transition-rate matrix (also known as a Q-matrix, [1] intensity matrix, [2] or infinitesimal generator matrix [3]) is an array of numbers describing the instantaneous rate at which a continuous-time Markov chain transitions between states.

  5. Markov chains on a measurable state space - Wikipedia

    en.wikipedia.org/wiki/Markov_chains_on_a...

    In 1953 the term Markov chain was used for stochastic processes with discrete or continuous index set, living on a countable or finite state space, see Doob. [1] or Chung. [2] Since the late 20th century it became more popular to consider a Markov chain as a stochastic process with discrete index set, living on a measurable state space. [3] [4] [5]

  6. Markovian arrival process - Wikipedia

    en.wikipedia.org/wiki/Markovian_arrival_process

    The Markov-modulated Poisson process or MMPP where m Poisson processes are switched between by an underlying continuous-time Markov chain. [8] If each of the m Poisson processes has rate λ i and the modulating continuous-time Markov has m × m transition rate matrix R , then the MAP representation is

  7. Continuous-time Markov chain - Wikipedia

    en.wikipedia.org/wiki/Continuous-time_Markov_chain

    Another discrete-time process that may be derived from a continuous-time Markov chain is a δ-skeleton—the (discrete-time) Markov chain formed by observing X(t) at intervals of δ units of time. The random variables X (0), X (δ), X (2δ), ... give the sequence of states visited by the δ-skeleton.

  8. Matrix analytic method - Wikipedia

    en.wikipedia.org/wiki/Matrix_analytic_method

    [1] [2] Such models are often described as M/G/1 type Markov chains because they can describe transitions in an M/G/1 queue. [3] [4] The method is a more complicated version of the matrix geometric method and is the classical solution method for M/G/1 chains. [5]

  9. Kolmogorov's criterion - Wikipedia

    en.wikipedia.org/wiki/Kolmogorov's_criterion

    Consider this figure depicting a section of a Markov chain with states i, j, k and l and the corresponding transition probabilities. Here Kolmogorov's criterion implies that the product of probabilities when traversing through any closed loop must be equal, so the product around the loop i to j to l to k returning to i must be equal to the loop the other way round,