enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Stopping time - Wikipedia

    en.wikipedia.org/wiki/Stopping_time

    Example of a stopping time: a hitting time of Brownian motion.The process starts at 0 and is stopped as soon as it hits 1. In probability theory, in particular in the study of stochastic processes, a stopping time (also Markov time, Markov moment, optional stopping time or optional time [1]) is a specific type of “random time”: a random variable whose value is interpreted as the time at ...

  3. Markov property - Wikipedia

    en.wikipedia.org/wiki/Markov_property

    The term strong Markov property is similar to the Markov property, except that the meaning of "present" is defined in terms of a random variable known as a stopping time. The term Markov assumption is used to describe a model where the Markov property is assumed to hold, such as a hidden Markov model .

  4. Markov chain - Wikipedia

    en.wikipedia.org/wiki/Markov_chain

    Markov chains and continuous-time Markov processes are useful in chemistry when physical systems closely approximate the Markov property. For example, imagine a large number n of molecules in solution in state A, each of which can undergo a chemical reaction to state B with a certain average rate. Perhaps the molecule is an enzyme, and the ...

  5. Markov Chains and Mixing Times - Wikipedia

    en.wikipedia.org/wiki/Markov_Chains_and_Mixing_Times

    A family of Markov chains is said to be rapidly mixing if the mixing time is a polynomial function of some size parameter of the Markov chain, and slowly mixing otherwise. This book is about finite Markov chains, their stationary distributions and mixing times, and methods for determining whether Markov chains are rapidly or slowly mixing. [1] [4]

  6. Markov chain mixing time - Wikipedia

    en.wikipedia.org/wiki/Markov_chain_mixing_time

    In probability theory, the mixing time of a Markov chain is the time until the Markov chain is "close" to its steady state distribution.. More precisely, a fundamental result about Markov chains is that a finite state irreducible aperiodic chain has a unique stationary distribution π and, regardless of the initial state, the time-t distribution of the chain converges to π as t tends to infinity.

  7. Examples of Markov chains - Wikipedia

    en.wikipedia.org/wiki/Examples_of_Markov_chains

    A game of snakes and ladders or any other game whose moves are determined entirely by dice is a Markov chain, indeed, an absorbing Markov chain. This is in contrast to card games such as blackjack, where the cards represent a 'memory' of the past moves. To see the difference, consider the probability for a certain event in the game.

  8. Markov model - Wikipedia

    en.wikipedia.org/wiki/Markov_model

    The simplest Markov model is the Markov chain.It models the state of a system with a random variable that changes through time. In this context, the Markov property indicates that the distribution for this variable depends only on the distribution of a previous state.

  9. Markov renewal process - Wikipedia

    en.wikipedia.org/wiki/Markov_renewal_process

    A semi-Markov process (defined in the above bullet point) in which all the holding times are exponentially distributed is called a continuous-time Markov chain. In other words, if the inter-arrival times are exponentially distributed and if the waiting time in a state and the next state reached are independent, we have a continuous-time Markov ...