Stochastic Markov Gradient Descent and Training Low-Bit Neural Networks

08/25/2020
by   Jonathan Ashbrock, et al.
0

The massive size of modern neural networks has motivated substantial recent interest in neural network quantization. We introduce Stochastic Markov Gradient Descent (SMGD), a discrete optimization method applicable to training quantized neural networks. The SMGD algorithm is designed for settings where memory is highly constrained during training. We provide theoretical guarantees of algorithm performance as well as encouraging numerical results.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset