enow.com Web Search

Search results

  1. Results from the WOW.Com Content Network
  2. Heat equation - Wikipedia

    en.wikipedia.org/wiki/Heat_equation

    The steady-state heat equation for a volume that contains a heat source (the inhomogeneous case), is the Poisson's equation: − k ∇ 2 u = q {\displaystyle -k\nabla ^{2}u=q} where u is the temperature , k is the thermal conductivity and q is the rate of heat generation per unit volume.

  3. Markov chain - Wikipedia

    en.wikipedia.org/wiki/Markov_chain

    Markov chain models have been used in advanced baseball analysis since 1960, although their use is still rare. Each half-inning of a baseball game fits the Markov chain state when the number of runners and outs are considered. During any at-bat, there are 24 possible combinations of number of outs and position of the runners.

  4. Examples of Markov chains - Wikipedia

    en.wikipedia.org/wiki/Examples_of_Markov_chains

    A game of snakes and ladders or any other game whose moves are determined entirely by dice is a Markov chain, indeed, an absorbing Markov chain. This is in contrast to card games such as blackjack, where the cards represent a 'memory' of the past moves. To see the difference, consider the probability for a certain event in the game.

  5. Master equation - Wikipedia

    en.wikipedia.org/wiki/Master_equation

    A master equation may be used to model a set of chemical reactions when the number of molecules of one or more species is small (of the order of 100 or 1000 molecules). [4] The chemical master equation can also solved for the very large models, such as the DNA damage signal from fungal pathogen Candida albicans. [5]

  6. Matrix analytic method - Wikipedia

    en.wikipedia.org/wiki/Matrix_analytic_method

    In probability theory, the matrix analytic method is a technique to compute the stationary probability distribution of a Markov chain which has a repeating structure (after some point) and a state space which grows unboundedly in no more than one dimension.

  7. Markov Chains and Mixing Times - Wikipedia

    en.wikipedia.org/wiki/Markov_Chains_and_Mixing_Times

    The mixing time of a Markov chain is the number of steps needed for this convergence to happen, to a suitable degree of accuracy. A family of Markov chains is said to be rapidly mixing if the mixing time is a polynomial function of some size parameter of the Markov chain, and slowly mixing otherwise. This book is about finite Markov chains ...

  8. Markov renewal process - Wikipedia

    en.wikipedia.org/wiki/Markov_renewal_process

    A semi-Markov process (defined in the above bullet point) in which all the holding times are exponentially distributed is called a continuous-time Markov chain. In other words, if the inter-arrival times are exponentially distributed and if the waiting time in a state and the next state reached are independent, we have a continuous-time Markov ...

  9. Detailed balance - Wikipedia

    en.wikipedia.org/wiki/Detailed_balance

    A Markov process is called a reversible Markov process or reversible Markov chain if there exists a positive stationary distribution π that satisfies the detailed balance equations [13] =, where P ij is the Markov transition probability from state i to state j, i.e. P ij = P(X t = j | X t − 1 = i), and π i and π j are the equilibrium probabilities of being in states i and j, respectively ...