Skip to yearly menu bar Skip to main content


Poster

Pessimism for Offline Linear Contextual Bandits using p Confidence Sets

Gene Li · Cong Ma · Nati Srebro

Hall J (level 1) #823

Keywords: [ offline reinforcement learning ] [ pessimism ] [ linear contextual bandits ]


Abstract: We present a family {π^p}p1 of pessimistic learning rules for offline learning of linear contextual bandits, relying on confidence sets with respect to different p norms, where π^2 corresponds to Bellman-consistent pessimism (BCP), while π^ is a novel generalization of lower confidence bound (LCB) to the linear setting. We show that the novel π^ learning rule is, in a sense, adaptively optimal, as it achieves the minimax performance (up to log factors) against all q-constrained problems, and as such it strictly dominates all other predictors in the family, including π^2.

Chat is not available.