Asymptotically efficient one-step stochastic gradient descent

06/09/2023
by   Alain Bensoussan, et al.
0

A generic, fast and asymptotically efficient method for parametric estimation is described. It is based on the stochastic gradient descent on the loglikelihood function corrected by a single step of the Fisher scoring algorithm. We show theoretically and by simulations in the i.i.d. setting that it is an interesting alternative to the usual stochastic gradient descent with averaging or the adaptative stochastic gradient descent.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset