How To Make the Gradients Small Stochastically

01/08/2018
by   Zeyuan Allen-Zhu, et al.
0

In convex stochastic optimization, convergence rates in terms of minimizing the objective have been well-established. However, in terms of making the gradients small, the best known convergence rate was O(ε^-8/3) and it was left open how to improve it. In this paper, we improve this rate to Õ(ε^-2), which is optimal up to log factors.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset