Toward Better PAC-Bayes Bounds for Uniformly Stable Algorithms

Sijia Zhou · Yunwen Lei · Ata Kaban

Great Hall & Hall B1+B2 (level 1) #1716
[ ]
Wed 13 Dec 8:45 a.m. PST — 10:45 a.m. PST

Abstract: We give sharper bounds for uniformly stable randomized algorithms in a PAC-Bayesian framework, which improve the existing results by up to a factor of $\sqrt{n}$ (ignoring a log factor), where $n$ is the sample size. The key idea is to bound the moment generating function of the generalization gap using concentration of weakly dependent random variables due to Bousquet et al (2020). We introduce an assumption of sub-exponential stability parameter, which allows a general treatment that we instantiate in two applications: stochastic gradient descent and randomized coordinate descent. Our results eliminate the requirement of strong convexity from previous results, and hold for non-smooth convex problems.

Chat is not available.