enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Matrix analytic method - Wikipedia

    en.wikipedia.org/wiki/Matrix_analytic_method

    [1] [2] Such models are often described as M/G/1 type Markov chains because they can describe transitions in an M/G/1 queue. [ 3 ] [ 4 ] The method is a more complicated version of the matrix geometric method and is the classical solution method for M/G/1 chains.

  3. Markov chain - Wikipedia

    en.wikipedia.org/wiki/Markov_chain

    Also, the growth (and composition) of copolymers may be modeled using Markov chains. Based on the reactivity ratios of the monomers that make up the growing polymer chain, the chain's composition may be calculated (for example, whether monomers tend to add in alternating fashion or in long runs of the same monomer).

  4. Markov Chains and Mixing Times - Wikipedia

    en.wikipedia.org/wiki/Markov_Chains_and_Mixing_Times

    A family of Markov chains is said to be rapidly mixing if the mixing time is a polynomial function of some size parameter of the Markov chain, and slowly mixing otherwise. This book is about finite Markov chains, their stationary distributions and mixing times, and methods for determining whether Markov chains are rapidly or slowly mixing. [1] [4]

  5. Examples of Markov chains - Wikipedia

    en.wikipedia.org/wiki/Examples_of_Markov_chains

    A game of snakes and ladders or any other game whose moves are determined entirely by dice is a Markov chain, indeed, an absorbing Markov chain. This is in contrast to card games such as blackjack, where the cards represent a 'memory' of the past moves. To see the difference, consider the probability for a certain event in the game.

  6. Euler–Maruyama method - Wikipedia

    en.wikipedia.org/wiki/Euler–Maruyama_method

    with initial condition X 0 = x 0, where W t denotes the Wiener process, and suppose that we wish to solve this SDE on some interval of time [0, T]. Then the Euler–Maruyama approximation to the true solution X is the Markov chain Y defined as follows: Partition the interval [0, T] into N equal subintervals of width >:

  7. Continuous-time Markov chain - Wikipedia

    en.wikipedia.org/wiki/Continuous-time_Markov_chain

    Another discrete-time process that may be derived from a continuous-time Markov chain is a δ-skeleton—the (discrete-time) Markov chain formed by observing X(t) at intervals of δ units of time. The random variables X (0), X (δ), X (2δ), ... give the sequence of states visited by the δ-skeleton.

  8. Discrete Poisson equation - Wikipedia

    en.wikipedia.org/wiki/Discrete_Poisson_equation

    The discrete Poisson's equation arises in the theory of Markov chains. It appears as the relative value function for the dynamic programming equation in a Markov decision process, and as the control variate for application in simulation variance reduction. [6] [7] [8]

  9. Uniformization (probability theory) - Wikipedia

    en.wikipedia.org/wiki/Uniformization...

    In probability theory, uniformization method, (also known as Jensen's method [1] or the randomization method [2]) is a method to compute transient solutions of finite state continuous-time Markov chains, by approximating the process by a discrete-time Markov chain. [2]