Timezone: »

Generalization Bounds for Stochastic Gradient Descent via Localized $\varepsilon$-Covers
Sejun Park · Umut Simsekli · Murat Erdogdu

Wed Nov 30 09:00 AM -- 11:00 AM (PST) @ Hall J #309
In this paper, we propose a new covering technique localized for the trajectories of SGD. This localization provides an algorithm-specific complexity measured by the covering number, which can have dimension-independent cardinality in contrast to standard uniform covering arguments that result in exponential dimension dependency. Based on this localized construction, we show that if the objective function is a finite perturbation of a piecewise strongly convex and smooth function with $P$ pieces, i.e., non-convex and non-smooth in general, the generalization error can be upper bounded by $O(\sqrt{(\log n\log(nP))/n})$, where $n$ is the number of data samples. In particular, this rate is independent of dimension and does not require early stopping and decaying step size. Finally, we employ these results in various contexts and derive generalization bounds for multi-index linear models, multi-class support vector machines, and $K$-means clustering for both hard and soft label setups, improving the previously known state-of-the-art rates.

Author Information

Sejun Park (Korea University)
Umut Simsekli (Inria Paris / ENS)
Murat Erdogdu (University of Toronto)

More from the Same Authors