Minimax Learning of Ergodic Markov Chains

Geoffrey Wolfer, Aryeh Kontorovich

Research output: Contribution to journalConference articlepeer-review

16 Scopus citations

Abstract

We compute the finite-sample minimax (modulo logarithmic factors) sample complexity of learning the parameters of a finite Markov chain from a single long sequence of states. Our error metric is a natural variant of total variation. The sample complexity necessarily depends on the spectral gap and minimal stationary probability of the unknown chain, for which there are known finite-sample estimators with fully empirical confidence intervals. To our knowledge, this is the first PAC-type result with nearly matching (up to logarithmic factors) upper and lower bounds for learning, in any metric, in the context of Markov chains.

Original languageEnglish
Pages (from-to)904-930
Number of pages27
JournalProceedings of Machine Learning Research
Volume98
StatePublished - 1 Jan 2019
Event30th International Conference on Algorithmic Learning Theory, ALT 2019 - Chicago, United States
Duration: 22 Mar 201924 Mar 2019

Keywords

  • ergodic Markov chain
  • learning
  • minimax

ASJC Scopus subject areas

  • Artificial Intelligence
  • Software
  • Control and Systems Engineering
  • Statistics and Probability

Fingerprint

Dive into the research topics of 'Minimax Learning of Ergodic Markov Chains'. Together they form a unique fingerprint.

Cite this