Abstract:
We consider the framework of non-stationary stochastic optimization [Besbes et.al. 2015] with squared error losses and noisy gradient feedback where the dynamic regret of an online learner against a time varying comparator sequence is studied. Motivated from the theory of non-parametric regression, we introduce a \emph{new variational constraint} that enforces the comparator sequence to belong to a discrete order Total Variation ball of radius . This variational constraint models comparators that have piecewise polynomial structure which has many relevant practical applications [Tibshirani2015]. By establishing connections to the theory of wavelet based non-parametric regression, we design a \emph{polynomial time} algorithm that achieves the nearly \emph{optimal dynamic regret} of . The proposed policy is \emph{adaptive to the unknown radius} . Further, we show that the same policy is minimax optimal for several other non-parametric families of interest.
Chat is not available.