Poster
Learning Multiple Markov Chains via Adaptive Allocation
Mohammad Sadegh Talebi · Odalric-Ambrym Maillard
East Exhibition Hall B, C #14
Keywords: [ Bandit Algorithms ] [ Algorithms ] [ Exploration ] [ Algorithms -> Active Learning; Algorithms -> Online Learning; Reinforcement Learning and Planning ]
We study the problem of learning the transition matrices of a set of Markov chains from a single stream of observations on each chain. We assume that the Markov chains are ergodic but otherwise unknown. The learner can sample Markov chains sequentially to observe their states. The goal of the learner is to sequentially select various chains to learn transition matrices uniformly well with respect to some loss function. We introduce a notion of loss that naturally extends the squared loss for learning distributions to the case of Markov chains, and further characterize the notion of being \emph{uniformly good} in all problem instances. We present a novel learning algorithm that efficiently balances \emph{exploration} and \emph{exploitation} intrinsic to this problem, without any prior knowledge of the chains. We provide finite-sample PAC-type guarantees on the performance of the algorithm. Further, we show that our algorithm asymptotically attains an optimal loss.
Live content is unavailable. Log in and register to view live content