markov_chain
1. [ noun ] (simulation) a Markov process for which the parameter is discrete time values
Synonyms: markoff_chain
Related terms: markoff_process
Similar spelling:   markoff_chain
  Markovich
  markovian
  mark_of_Cain