markoff_process
1. [ noun ] (simulation) a simple stochastic process in which the distribution of future states depends only on the present state and not on how it arrived in the present state
Synonyms: markov_process
Related terms: stochastic_process markoff_chain
Similar spelling:   markov_process
  markoff_chain