Part of Advances in Neural Information Processing Systems 31 (NeurIPS 2018)
Zeyuan Allen-Zhu
Stochastic gradient descent (SGD) gives an optimal convergence rate when minimizing convex stochastic objectives f(x). However, in terms of making the gradients small, the original SGD does not give an optimal rate, even when f(x) is convex. If f(x) is convex, to find a point with gradient norm ε, we design an algorithm SGD3 with a near-optimal rate ˜O(ε−2), improving the best known rate O(ε−8/3). If f(x) is nonconvex, to find its ε-approximate local minimum, we design an algorithm SGD5 with rate ˜O(ε−3.5), where previously SGD variants only achieve ˜O(ε−4). This is no slower than the best known stochastic version of Newton's method in all parameter regimes.