Part of Advances in Neural Information Processing Systems 31 (NeurIPS 2018)
Zeyuan Allen-Zhu
We design a stochastic algorithm to find ε-approximate local minima of any smooth nonconvex function in rate O(ε−3.25), with only oracle access to stochastic gradients. The best result before this work was O(ε−4) by stochastic gradient descent (SGD).