Markov Chains

Medical Dictionary -> Markov Chains

Markov Chains


A stochastic process such that the conditional probability distribution for a state at any future instant, given the present state, is unaffected by any additional knowledge of the past history of the system.


© MedicalDictionaryweb.com 2012 | Contact Us | Terms of Use | Teeth Whitening | Low Carb Foods and Diets