Poster
RSN: Randomized Subspace Newton
Robert Gower · Dmitry Kovalev · Felix Lieder · Peter Richtarik
East Exhibition Hall B, C #131
Keywords: [ Convex Optimization ] [ Optimization ] [ Stochastic Optimization ]
We develop a randomized Newton method capable of solving learning problems with huge dimensional feature spaces, which is a common setting in applications such as medical imaging, genomics and seismology. Our method leverages randomized sketching in a new way, by finding the Newton direction constrained to the space spanned by a random sketch. We develop a simple global linear convergence theory that holds for practically all sketching techniques, which gives the practitioners the freedom to design custom sketching approaches suitable for particular applications. We perform numerical experiments which demonstrate the efficiency of our method as compared to accelerated gradient descent and the full Newton method. Our method can be seen as a refinement and a randomized extension of the results of Karimireddy, Stich, and Jaggi (2019).
Live content is unavailable. Log in and register to view live content