Timezone: »

SGD Algorithms based on Incomplete U-statistics: Large-Scale Minimization of Empirical Risk
Guillaume Papa · Stéphan Clémençon · Aurélien Bellet

Tue Dec 08 04:00 PM -- 08:59 PM (PST) @ 210 C #75

In many learning problems, ranging from clustering to ranking through metric learning, empirical estimates of the risk functional consist of an average over tuples (e.g., pairs or triplets) of observations, rather than over individual observations. In this paper, we focus on how to best implement a stochastic approximation approach to solve such risk minimization problems. We argue that in the large-scale setting, gradient estimates should be obtained by sampling tuples of data points with replacement (incomplete U-statistics) instead of sampling data points without replacement (complete U-statistics based on subsamples). We develop a theoretical framework accounting for the substantial impact of this strategy on the generalization ability of the prediction model returned by the Stochastic Gradient Descent (SGD) algorithm. It reveals that the method we promote achieves a much better trade-off between statistical accuracy and computational cost. Beyond the rate bound analysis, experiments on AUC maximization and metric learning provide strong empirical evidence of the superiority of the proposed approach.

Author Information

Guillaume Papa (Telecom paristech)
Stéphan Clémençon (Telecom ParisTech)
Aurélien Bellet (Telecom ParisTech)

More from the Same Authors