enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Stopping time - Wikipedia

    en.wikipedia.org/wiki/Stopping_time

    Example of a stopping time: a hitting time of Brownian motion.The process starts at 0 and is stopped as soon as it hits 1. In probability theory, in particular in the study of stochastic processes, a stopping time (also Markov time, Markov moment, optional stopping time or optional time [1]) is a specific type of “random time”: a random variable whose value is interpreted as the time at ...

  3. Markov Chains and Mixing Times - Wikipedia

    en.wikipedia.org/wiki/Markov_Chains_and_Mixing_Times

    A family of Markov chains is said to be rapidly mixing if the mixing time is a polynomial function of some size parameter of the Markov chain, and slowly mixing otherwise. This book is about finite Markov chains, their stationary distributions and mixing times, and methods for determining whether Markov chains are rapidly or slowly mixing. [1] [4]

  4. Markov property - Wikipedia

    en.wikipedia.org/wiki/Markov_property

    A Markov random field extends this property to two or more dimensions or to random variables defined for an interconnected network of items. [1] An example of a model for such a field is the Ising model. A discrete-time stochastic process satisfying the Markov property is known as a Markov chain.

  5. Markov chain - Wikipedia

    en.wikipedia.org/wiki/Markov_chain

    Markov chains and continuous-time Markov processes are useful in chemistry when physical systems closely approximate the Markov property. For example, imagine a large number n of molecules in solution in state A, each of which can undergo a chemical reaction to state B with a certain average rate. Perhaps the molecule is an enzyme, and the ...

  6. Itô diffusion - Wikipedia

    en.wikipedia.org/wiki/Itô_diffusion

    The strong Markov property is a generalization of the Markov property above in which t is replaced by a suitable random time τ : Ω → [0, +∞] known as a stopping time. So, for example, rather than "restarting" the process X at time t = 1, one could "restart" whenever X first reaches some specified point p of R n.

  7. Examples of Markov chains - Wikipedia

    en.wikipedia.org/wiki/Examples_of_Markov_chains

    A game of snakes and ladders or any other game whose moves are determined entirely by dice is a Markov chain, indeed, an absorbing Markov chain. This is in contrast to card games such as blackjack, where the cards represent a 'memory' of the past moves. To see the difference, consider the probability for a certain event in the game.

  8. Optional stopping theorem - Wikipedia

    en.wikipedia.org/wiki/Optional_stopping_theorem

    Suppose further that the walk stops if it reaches 0 or m ≥ a; the time at which this first occurs is a stopping time. If it is known that the expected time at which the walk ends is finite (say, from Markov chain theory), the optional stopping theorem predicts that the expected stop position is equal to the initial position a.

  9. Continuous-time Markov chain - Wikipedia

    en.wikipedia.org/wiki/Continuous-time_Markov_chain

    A continuous-time Markov chain (CTMC) is a continuous stochastic process in which, for each state, the process will change state according to an exponential random variable and then move to a different state as specified by the probabilities of a stochastic matrix.