Skip to yearly menu bar Skip to main content


Poster

Natasha 2: Faster Non-Convex Optimization Than SGD

Zeyuan Allen-Zhu

Room 210 #50

Keywords: [ Learning Theory ] [ Non-Convex Optimization ]


Abstract: We design a stochastic algorithm to find ε-approximate local minima of any smooth nonconvex function in rate O(ε3.25), with only oracle access to stochastic gradients. The best result before this work was O(ε4) by stochastic gradient descent (SGD).

Live content is unavailable. Log in and register to view live content