fbpx
Wikipedia

Stochastic matrix

In mathematics, a stochastic matrix is a square matrix used to describe the transitions of a Markov chain. Each of its entries is a nonnegative real number representing a probability.[1][2]: 9–11  It is also called a probability matrix, transition matrix, substitution matrix, or Markov matrix.[2]: 9–11  The stochastic matrix was first developed by Andrey Markov at the beginning of the 20th century, and has found use throughout a wide variety of scientific fields, including probability theory, statistics, mathematical finance and linear algebra, as well as computer science and population genetics.[2]: 1–8  There are several different definitions and types of stochastic matrices:[2]: 9–11 

A right stochastic matrix is a real square matrix, with each row summing to 1.
A left stochastic matrix is a real square matrix, with each column summing to 1.
A doubly stochastic matrix is a square matrix of nonnegative real numbers with each row and column summing to 1.

In the same vein, one may define a stochastic vector (also called probability vector) as a vector whose elements are nonnegative real numbers which sum to 1. Thus, each row of a right stochastic matrix (or column of a left stochastic matrix) is a stochastic vector.[2]: 9–11  A common convention in English language mathematics literature is to use row vectors of probabilities and right stochastic matrices rather than column vectors of probabilities and left stochastic matrices; this article follows that convention.[2]: 1–8  In addition, a substochastic matrix is a real square matrix whose row sums are all

History edit

 
Andrey Markov in 1886

The stochastic matrix was developed alongside the Markov chain by Andrey Markov, a Russian mathematician and professor at St. Petersburg University who first published on the topic in 1906.[2]: 1–8 [3] His initial intended uses were for linguistic analysis and other mathematical subjects like card shuffling, but both Markov chains and matrices rapidly found use in other fields.[2]: 1–8 [3][4]

Stochastic matrices were further developed by scholars such as Andrey Kolmogorov, who expanded their possibilities by allowing for continuous-time Markov processes.[5] By the 1950s, articles using stochastic matrices had appeared in the fields of econometrics[6] and circuit theory.[7] In the 1960s, stochastic matrices appeared in an even wider variety of scientific works, from behavioral science[8] to geology[9][10] to residential planning.[11] In addition, much mathematical work was also done through these decades to improve the range of uses and functionality of the stochastic matrix and Markovian processes more generally.

From the 1970s to present, stochastic matrices have found use in almost every field that requires formal analysis, from structural science[12] to medical diagnosis[13] to personnel management.[14] In addition, stochastic matrices have found wide use in land change modeling, usually under the term Markov matrix.[15]

Definition and properties edit

A stochastic matrix describes a Markov chain Xt over a finite state space S with cardinality α.

If the probability of moving from j to i in one time step is Pr(j|i) = Pi,j, the stochastic matrix P is given by using Pi,j as the i-th row and j-th column element, e.g.,

 

Since the total of transition probability from a state i to all other states must be 1,

 

thus this matrix is a right stochastic matrix.[2]: 1–8 

The above elementwise sum across each row i of P may be more concisely written as P1 = 1, where 1 is the α-dimensional column vector of all ones. Using this, it can be seen that the product of two right stochastic matrices P and P′′ is also right stochastic: PP′′ 1 = P′ (P′′ 1) = P1 = 1. In general, the k-th power Pk of a right stochastic matrix P is also right stochastic. The probability of transitioning from i to j in two steps is then given by the (i, j)-th element of the square of P:

 

In general, the probability transition of going from any state to another state in a finite Markov chain given by the matrix P in k steps is given by Pk.

An initial probability distribution of states, specifying where the system might be initially and with what probabilities, is given as a row vector.

A stationary probability vector π is defined as a distribution, written as a row vector, that does not change under application of the transition matrix; that is, it is defined as a probability distribution on the set {1, …, n} which is also a row eigenvector of the probability matrix, associated with eigenvalue 1:

 

It can be shown that the spectral radius of any stochastic matrix is one. By the Gershgorin circle theorem, all of the eigenvalues of a stochastic matrix have absolute values less than or equal to one. Additionally, every right stochastic matrix has an "obvious" column eigenvector associated to the eigenvalue 1: the vector 1 used above, whose coordinates are all equal to 1. As left and right eigenvalues of a square matrix are the same, every stochastic matrix has, at least, a row eigenvector associated to the eigenvalue 1 and the largest absolute value of all its eigenvalues is also 1. Finally, the Brouwer Fixed Point Theorem (applied to the compact convex set of all probability distributions of the finite set {1, …, n}) implies that there is some left eigenvector which is also a stationary probability vector.

On the other hand, the Perron–Frobenius theorem also ensures that every irreducible stochastic matrix has such a stationary vector, and that the largest absolute value of an eigenvalue is always 1. However, this theorem cannot be applied directly to such matrices because they need not be irreducible.

In general, there may be several such vectors. However, for a matrix with strictly positive entries (or, more generally, for an irreducible aperiodic stochastic matrix), this vector is unique and can be computed by observing that for any i we have the following limit,

 

where πj is the j-th element of the row vector π. Among other things, this says that the long-term probability of being in a state j is independent of the initial state i. That both of these computations give the same stationary vector is a form of an ergodic theorem, which is generally true in a wide variety of dissipative dynamical systems: the system evolves, over time, to a stationary state.

Intuitively, a stochastic matrix represents a Markov chain; the application of the stochastic matrix to a probability distribution redistributes the probability mass of the original distribution while preserving its total mass. If this process is applied repeatedly, the distribution converges to a stationary distribution for the Markov chain.[2]: 55–59 

Example: Cat and Mouse edit

Suppose there is a timer and a row of five adjacent boxes. At time zero, a cat is in the first box, and a mouse is in the fifth box. The cat and the mouse both jump to a random adjacent box when the timer advances. For example, if the cat is in the second box and the mouse is in the fourth, the probability that the cat will be in the first box and the mouse in the fifth after the timer advances is one fourth. If the cat is in the first box and the mouse is in the fifth, the probability that the cat will be in box two and the mouse will be in box four after the timer advances is one. The cat eats the mouse if both end up in the same box, at which time the game ends. Let the random variable K be the time the mouse stays in the game.

The Markov chain that represents this game contains the following five states specified by the combination of positions (cat,mouse). Note that while a naive enumeration of states would list 25 states, many are impossible either because the mouse can never have a lower index than the cat (as that would mean the mouse occupied the cat's box and survived to move past it), or because the sum of the two indices will always have even parity. In addition, the 3 possible states that lead to the mouse's death are combined into one:

  • State 1: (1,3)
  • State 2: (1,5)
  • State 3: (2,4)
  • State 4: (3,5)
  • State 5: game over: (2,2), (3,3) & (4,4).

We use a stochastic matrix,   (below), to represent the transition probabilities of this system (rows and columns in this matrix are indexed by the possible states listed above, with the pre-transition state as the row and post-transition state as the column).[2]: 1–8  For instance, starting from state 1 – 1st row – it is impossible for the system to stay in this state, so  ; the system also cannot transition to state 2 – because the cat would have stayed in the same box – so  , and by a similar argument for the mouse,  . Transitions to states 3 or 5 are allowed, and thus   .

 

Long-term averages edit

No matter what the initial state, the cat will eventually catch the mouse (with probability 1) and a stationary state π = (0,0,0,0,1) is approached as a limit.[2]: 55–59  To compute the long-term average or expected value of a stochastic variable  , for each state   and time   there is a contribution of  . Survival can be treated as a binary variable with   for a surviving state and   for the terminated state. The states with   do not contribute to the long-term average.

Phase-type representation edit

 
The survival function of the mouse. The mouse will survive at least the first time step.

As State 5 is an absorbing state, the distribution of time to absorption is discrete phase-type distributed. Suppose the system starts in state 2, represented by the vector  . The states where the mouse has perished don't contribute to the survival average so state five can be ignored. The initial state and transition matrix can be reduced to,

 

and

 

where   is the identity matrix, and   represents a column matrix of all ones that acts as a sum over states.

Since each state is occupied for one step of time the expected time of the mouse's survival is just the sum of the probability of occupation over all surviving states and steps in time,

 

Higher order moments are given by

 

See also edit

References edit

  1. ^ Asmussen, S. R. (2003). "Markov Chains". Applied Probability and Queues. Stochastic Modelling and Applied Probability. Vol. 51. pp. 3–8. doi:10.1007/0-387-21525-5_1. ISBN 978-0-387-00211-8.
  2. ^ a b c d e f g h i j k l Gagniuc, Paul A. (2017). Markov Chains: From Theory to Implementation and Experimentation. USA, NJ: John Wiley & Sons. pp. 9–11. ISBN 978-1-119-38755-8.
  3. ^ a b Hayes, Brian (2013). "First links in the Markov chain". American Scientist. 101 (2): 92–96. doi:10.1511/2013.101.92.
  4. ^ Charles Miller Grinstead; James Laurie Snell (1997). Introduction to Probability. American Mathematical Soc. pp. 464–466. ISBN 978-0-8218-0749-1.
  5. ^ Kendall, D. G.; Batchelor, G. K.; Bingham, N. H.; Hayman, W. K.; Hyland, J. M. E.; Lorentz, G. G.; Moffatt, H. K.; Parry, W.; Razborov, A. A.; Robinson, C. A.; Whittle, P. (1990). "Andrei Nikolaevich Kolmogorov (1903–1987)". Bulletin of the London Mathematical Society. 22 (1): 33. doi:10.1112/blms/22.1.31.
  6. ^ Solow, Robert (1 January 1952). "On the Structure of Linear Models". Econometrica. 20 (1): 29–46. doi:10.2307/1907805. JSTOR 1907805.
  7. ^ Sittler, R. (1 December 1956). "Systems Analysis of Discrete Markov Processes". IRE Transactions on Circuit Theory. 3 (4): 257–266. doi:10.1109/TCT.1956.1086324. ISSN 0096-2007.
  8. ^ Evans, Selby (1 July 1967). "Vargus 7: Computed patterns from markov processes". Behavioral Science. 12 (4): 323–328. doi:10.1002/bs.3830120407. ISSN 1099-1743.
  9. ^ Gingerich, P. D. (1 January 1969). "Markov analysis of cyclic alluvial sediments". Journal of Sedimentary Research. 39 (1): 330–332. Bibcode:1969JSedR..39..330G. doi:10.1306/74d71c4e-2b21-11d7-8648000102c1865d. ISSN 1527-1404.
  10. ^ Krumbein, W. C.; Dacey, Michael F. (1 March 1969). "Markov chains and embedded Markov chains in geology". Journal of the International Association for Mathematical Geology. 1 (1): 79–96. doi:10.1007/BF02047072. ISSN 0020-5958.
  11. ^ Wolfe, Harry B. (1 May 1967). "Models for Conditioning Aging of Residential Structures". Journal of the American Institute of Planners. 33 (3): 192–196. doi:10.1080/01944366708977915. ISSN 0002-8991.
  12. ^ Krenk, S. (November 1989). "A Markov matrix for fatigue load simulation and rainflow range evaluation". Structural Safety. 6 (2–4): 247–258. doi:10.1016/0167-4730(89)90025-8.
  13. ^ Beck, J.Robert; Pauker, Stephen G. (1 December 1983). "The Markov Process in Medical Prognosis". Medical Decision Making. 3 (4): 419–458. doi:10.1177/0272989X8300300403. ISSN 0272-989X. PMID 6668990.
  14. ^ Gotz, Glenn A.; McCall, John J. (1 March 1983). "Sequential Analysis of the Stay/Leave Decision: U.S. Air Force Officers". Management Science. 29 (3): 335–351. doi:10.1287/mnsc.29.3.335. ISSN 0025-1909.
  15. ^ Kamusoko, Courage; Aniya, Masamu; Adi, Bongo; Manjoro, Munyaradzi (1 July 2009). "Rural sustainability under threat in Zimbabwe – Simulation of future land use/cover changes in the Bindura district based on the Markov-cellular automata model". Applied Geography. 29 (3): 435–447. doi:10.1016/j.apgeog.2008.10.002.

stochastic, matrix, matrix, whose, elements, stochastic, random, matrix, mathematics, stochastic, matrix, square, matrix, used, describe, transitions, markov, chain, each, entries, nonnegative, real, number, representing, probability, also, called, probability. For a matrix whose elements are stochastic see Random matrix In mathematics a stochastic matrix is a square matrix used to describe the transitions of a Markov chain Each of its entries is a nonnegative real number representing a probability 1 2 9 11 It is also called a probability matrix transition matrix substitution matrix or Markov matrix 2 9 11 The stochastic matrix was first developed by Andrey Markov at the beginning of the 20th century and has found use throughout a wide variety of scientific fields including probability theory statistics mathematical finance and linear algebra as well as computer science and population genetics 2 1 8 There are several different definitions and types of stochastic matrices 2 9 11 A right stochastic matrix is a real square matrix with each row summing to 1 A left stochastic matrix is a real square matrix with each column summing to 1 A doubly stochastic matrix is a square matrix of nonnegative real numbers with each row and column summing to 1 In the same vein one may define a stochastic vector also called probability vector as a vector whose elements are nonnegative real numbers which sum to 1 Thus each row of a right stochastic matrix or column of a left stochastic matrix is a stochastic vector 2 9 11 A common convention in English language mathematics literature is to use row vectors of probabilities and right stochastic matrices rather than column vectors of probabilities and left stochastic matrices this article follows that convention 2 1 8 In addition a substochastic matrix is a real square matrix whose row sums are all 1 displaystyle leq 1 Contents 1 History 2 Definition and properties 3 Example Cat and Mouse 3 1 Long term averages 3 2 Phase type representation 4 See also 5 ReferencesHistory edit nbsp Andrey Markov in 1886The stochastic matrix was developed alongside the Markov chain by Andrey Markov a Russian mathematician and professor at St Petersburg University who first published on the topic in 1906 2 1 8 3 His initial intended uses were for linguistic analysis and other mathematical subjects like card shuffling but both Markov chains and matrices rapidly found use in other fields 2 1 8 3 4 Stochastic matrices were further developed by scholars such as Andrey Kolmogorov who expanded their possibilities by allowing for continuous time Markov processes 5 By the 1950s articles using stochastic matrices had appeared in the fields of econometrics 6 and circuit theory 7 In the 1960s stochastic matrices appeared in an even wider variety of scientific works from behavioral science 8 to geology 9 10 to residential planning 11 In addition much mathematical work was also done through these decades to improve the range of uses and functionality of the stochastic matrix and Markovian processes more generally From the 1970s to present stochastic matrices have found use in almost every field that requires formal analysis from structural science 12 to medical diagnosis 13 to personnel management 14 In addition stochastic matrices have found wide use in land change modeling usually under the term Markov matrix 15 Definition and properties editA stochastic matrix describes a Markov chain Xt over a finite state space S with cardinality a If the probability of moving from j to i in one time step is Pr j i Pi j the stochastic matrix P is given by using Pi j as the i th row and j th column element e g P P 1 1 P 1 2 P 1 j P 1 a P 2 1 P 2 2 P 2 j P 2 a P i 1 P i 2 P i j P i a P a 1 P a 2 P a j P a a displaystyle P left begin matrix P 1 1 amp P 1 2 amp dots amp P 1 j amp dots amp P 1 alpha P 2 1 amp P 2 2 amp dots amp P 2 j amp dots amp P 2 alpha vdots amp vdots amp ddots amp vdots amp ddots amp vdots P i 1 amp P i 2 amp dots amp P i j amp dots amp P i alpha vdots amp vdots amp ddots amp vdots amp ddots amp vdots P alpha 1 amp P alpha 2 amp dots amp P alpha j amp dots amp P alpha alpha end matrix right nbsp Since the total of transition probability from a state i to all other states must be 1 j 1 a P i j 1 displaystyle sum j 1 alpha P i j 1 nbsp thus this matrix is a right stochastic matrix 2 1 8 The above elementwise sum across each row i of P may be more concisely written as P1 1 where 1 is the a dimensional column vector of all ones Using this it can be seen that the product of two right stochastic matrices P and P is also right stochastic P P 1 P P 1 P 1 1 In general the k th power Pk of a right stochastic matrix P is also right stochastic The probability of transitioning from i to j in two steps is then given by the i j th element of the square of P P 2 i j displaystyle left P 2 right i j nbsp In general the probability transition of going from any state to another state in a finite Markov chain given by the matrix P in k steps is given by Pk An initial probability distribution of states specifying where the system might be initially and with what probabilities is given as a row vector A stationary probability vector p is defined as a distribution written as a row vector that does not change under application of the transition matrix that is it is defined as a probability distribution on the set 1 n which is also a row eigenvector of the probability matrix associated with eigenvalue 1 p P p displaystyle boldsymbol pi P boldsymbol pi nbsp It can be shown that the spectral radius of any stochastic matrix is one By the Gershgorin circle theorem all of the eigenvalues of a stochastic matrix have absolute values less than or equal to one Additionally every right stochastic matrix has an obvious column eigenvector associated to the eigenvalue 1 the vector 1 used above whose coordinates are all equal to 1 As left and right eigenvalues of a square matrix are the same every stochastic matrix has at least a row eigenvector associated to the eigenvalue 1 and the largest absolute value of all its eigenvalues is also 1 Finally the Brouwer Fixed Point Theorem applied to the compact convex set of all probability distributions of the finite set 1 n implies that there is some left eigenvector which is also a stationary probability vector On the other hand the Perron Frobenius theorem also ensures that every irreducible stochastic matrix has such a stationary vector and that the largest absolute value of an eigenvalue is always 1 However this theorem cannot be applied directly to such matrices because they need not be irreducible In general there may be several such vectors However for a matrix with strictly positive entries or more generally for an irreducible aperiodic stochastic matrix this vector is unique and can be computed by observing that for any i we have the following limit lim k P k i j p j displaystyle lim k rightarrow infty left P k right i j boldsymbol pi j nbsp where pj is the j th element of the row vector p Among other things this says that the long term probability of being in a state j is independent of the initial state i That both of these computations give the same stationary vector is a form of an ergodic theorem which is generally true in a wide variety of dissipative dynamical systems the system evolves over time to a stationary state Intuitively a stochastic matrix represents a Markov chain the application of the stochastic matrix to a probability distribution redistributes the probability mass of the original distribution while preserving its total mass If this process is applied repeatedly the distribution converges to a stationary distribution for the Markov chain 2 55 59 Example Cat and Mouse editSuppose there is a timer and a row of five adjacent boxes At time zero a cat is in the first box and a mouse is in the fifth box The cat and the mouse both jump to a random adjacent box when the timer advances For example if the cat is in the second box and the mouse is in the fourth the probability that the cat will be in the first box and the mouse in the fifth after the timer advances is one fourth If the cat is in the first box and the mouse is in the fifth the probability that the cat will be in box two and the mouse will be in box four after the timer advances is one The cat eats the mouse if both end up in the same box at which time the game ends Let the random variable K be the time the mouse stays in the game The Markov chain that represents this game contains the following five states specified by the combination of positions cat mouse Note that while a naive enumeration of states would list 25 states many are impossible either because the mouse can never have a lower index than the cat as that would mean the mouse occupied the cat s box and survived to move past it or because the sum of the two indices will always have even parity In addition the 3 possible states that lead to the mouse s death are combined into one State 1 1 3 State 2 1 5 State 3 2 4 State 4 3 5 State 5 game over 2 2 3 3 amp 4 4 We use a stochastic matrix P displaystyle P nbsp below to represent the transition probabilities of this system rows and columns in this matrix are indexed by the possible states listed above with the pre transition state as the row and post transition state as the column 2 1 8 For instance starting from state 1 1st row it is impossible for the system to stay in this state so P 11 0 displaystyle P 11 0 nbsp the system also cannot transition to state 2 because the cat would have stayed in the same box so P 12 0 displaystyle P 12 0 nbsp and by a similar argument for the mouse P 14 0 displaystyle P 14 0 nbsp Transitions to states 3 or 5 are allowed and thus P 13 P 15 0 displaystyle P 13 P 15 neq 0 nbsp P 0 0 1 2 0 1 2 0 0 1 0 0 1 4 1 4 0 1 4 1 4 0 0 1 2 0 1 2 0 0 0 0 1 displaystyle P begin bmatrix 0 amp 0 amp 1 2 amp 0 amp 1 2 0 amp 0 amp 1 amp 0 amp 0 1 4 amp 1 4 amp 0 amp 1 4 amp 1 4 0 amp 0 amp 1 2 amp 0 amp 1 2 0 amp 0 amp 0 amp 0 amp 1 end bmatrix nbsp Long term averages edit No matter what the initial state the cat will eventually catch the mouse with probability 1 and a stationary state p 0 0 0 0 1 is approached as a limit 2 55 59 To compute the long term average or expected value of a stochastic variable Y displaystyle Y nbsp for each state S j displaystyle S j nbsp and time t k displaystyle t k nbsp there is a contribution of Y j k P S S j t t k displaystyle Y j k cdot P S S j t t k nbsp Survival can be treated as a binary variable with Y 1 displaystyle Y 1 nbsp for a surviving state and Y 0 displaystyle Y 0 nbsp for the terminated state The states with Y 0 displaystyle Y 0 nbsp do not contribute to the long term average Phase type representation edit nbsp The survival function of the mouse The mouse will survive at least the first time step As State 5 is an absorbing state the distribution of time to absorption is discrete phase type distributed Suppose the system starts in state 2 represented by the vector 0 1 0 0 0 displaystyle 0 1 0 0 0 nbsp The states where the mouse has perished don t contribute to the survival average so state five can be ignored The initial state and transition matrix can be reduced to t 0 1 0 0 T 0 0 1 2 0 0 0 1 0 1 4 1 4 0 1 4 0 0 1 2 0 displaystyle boldsymbol tau 0 1 0 0 qquad T begin bmatrix 0 amp 0 amp frac 1 2 amp 0 0 amp 0 amp 1 amp 0 frac 1 4 amp frac 1 4 amp 0 amp frac 1 4 0 amp 0 amp frac 1 2 amp 0 end bmatrix nbsp and I T 1 1 2 75 4 5 3 5 2 75 displaystyle I T 1 boldsymbol 1 begin bmatrix 2 75 4 5 3 5 2 75 end bmatrix nbsp where I displaystyle I nbsp is the identity matrix and 1 displaystyle mathbf 1 nbsp represents a column matrix of all ones that acts as a sum over states Since each state is occupied for one step of time the expected time of the mouse s survival is just the sum of the probability of occupation over all surviving states and steps in time E K t I T T 2 1 t I T 1 1 4 5 displaystyle E K boldsymbol tau left I T T 2 cdots right boldsymbol 1 boldsymbol tau I T 1 boldsymbol 1 4 5 nbsp Higher order moments are given by E K K 1 K n 1 n t I T n T n 1 1 displaystyle E K K 1 dots K n 1 n boldsymbol tau I T n T n 1 mathbf 1 nbsp See also editDensity matrix Markov kernel the equivalent of a stochastic matrix over a continuous state space Matrix difference equation Models of DNA evolution Muirhead s inequality Probabilistic automaton Transition rate matrix used to generalize the stochastic matrix to continuous timeReferences edit Asmussen S R 2003 Markov Chains Applied Probability and Queues Stochastic Modelling and Applied Probability Vol 51 pp 3 8 doi 10 1007 0 387 21525 5 1 ISBN 978 0 387 00211 8 a b c d e f g h i j k l Gagniuc Paul A 2017 Markov Chains From Theory to Implementation and Experimentation USA NJ John Wiley amp Sons pp 9 11 ISBN 978 1 119 38755 8 a b Hayes Brian 2013 First links in the Markov chain American Scientist 101 2 92 96 doi 10 1511 2013 101 92 Charles Miller Grinstead James Laurie Snell 1997 Introduction to Probability American Mathematical Soc pp 464 466 ISBN 978 0 8218 0749 1 Kendall D G Batchelor G K Bingham N H Hayman W K Hyland J M E Lorentz G G Moffatt H K Parry W Razborov A A Robinson C A Whittle P 1990 Andrei Nikolaevich Kolmogorov 1903 1987 Bulletin of the London Mathematical Society 22 1 33 doi 10 1112 blms 22 1 31 Solow Robert 1 January 1952 On the Structure of Linear Models Econometrica 20 1 29 46 doi 10 2307 1907805 JSTOR 1907805 Sittler R 1 December 1956 Systems Analysis of Discrete Markov Processes IRE Transactions on Circuit Theory 3 4 257 266 doi 10 1109 TCT 1956 1086324 ISSN 0096 2007 Evans Selby 1 July 1967 Vargus 7 Computed patterns from markov processes Behavioral Science 12 4 323 328 doi 10 1002 bs 3830120407 ISSN 1099 1743 Gingerich P D 1 January 1969 Markov analysis of cyclic alluvial sediments Journal of Sedimentary Research 39 1 330 332 Bibcode 1969JSedR 39 330G doi 10 1306 74d71c4e 2b21 11d7 8648000102c1865d ISSN 1527 1404 Krumbein W C Dacey Michael F 1 March 1969 Markov chains and embedded Markov chains in geology Journal of the International Association for Mathematical Geology 1 1 79 96 doi 10 1007 BF02047072 ISSN 0020 5958 Wolfe Harry B 1 May 1967 Models for Conditioning Aging of Residential Structures Journal of the American Institute of Planners 33 3 192 196 doi 10 1080 01944366708977915 ISSN 0002 8991 Krenk S November 1989 A Markov matrix for fatigue load simulation and rainflow range evaluation Structural Safety 6 2 4 247 258 doi 10 1016 0167 4730 89 90025 8 Beck J Robert Pauker Stephen G 1 December 1983 The Markov Process in Medical Prognosis Medical Decision Making 3 4 419 458 doi 10 1177 0272989X8300300403 ISSN 0272 989X PMID 6668990 Gotz Glenn A McCall John J 1 March 1983 Sequential Analysis of the Stay Leave Decision U S Air Force Officers Management Science 29 3 335 351 doi 10 1287 mnsc 29 3 335 ISSN 0025 1909 Kamusoko Courage Aniya Masamu Adi Bongo Manjoro Munyaradzi 1 July 2009 Rural sustainability under threat in Zimbabwe Simulation of future land use cover changes in the Bindura district based on the Markov cellular automata model Applied Geography 29 3 435 447 doi 10 1016 j apgeog 2008 10 002 Retrieved from https en wikipedia org w index php title Stochastic matrix amp oldid 1210741703, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.