Poster
Refined Learning Bounds for Kernel and Approximate -Means
Yong Liu
Keywords: [ Kernel Methods ] [ Clustering ] [ Theory ]
Abstract:
Kernel -means is one of the most popular approaches to clustering and its theoretical properties have been investigated for decades. However, the existing state-of-the-art risk bounds are of order , which do not match with the stated lower bound in terms of , where is the number of clusters and is the size of the training set. In this paper, we study the statistical properties of kernel -means and Nystr\"{o}m-based kernel -means, and obtain optimal clustering risk bounds, which improve the existing risk bounds. Particularly, based on a refined upper bound of Rademacher complexity [21], we first derive an optimal risk bound of rate for empirical risk minimizer (ERM), and further extend it to general cases beyond ERM. Then, we analyze the statistical effect of computational approximations of Nystr\"{o}m kernel -means, and prove that it achieves the same statistical accuracy as the original kernel -means considering only Nystr\"{o}m landmark points. We further relax the restriction of landmark points from to under a mild condition. Finally, we validate the theoretical findings via numerical experiments.
Chat is not available.