DeepShift: Towards Multiplication-Less Neural Networks

05/30/2019
by   Mostafa Elhoushi, et al.
0

Deep learning models, especially DCNN have obtained high accuracies in several computer vision applications. However, for deployment in mobile environments, the high computation and power budget proves to be a major bottleneck. Convolution layers and fully connected layers, because of their intense use of multiplications, are the dominant contributer to this computation budget. This paper, proposes to tackle this problem by introducing two new operations: convolutional shifts and fully-connected shifts, that replace multiplications all together and use bitwise shift and bitwise negation instead. This family of neural network architectures (that use convolutional shifts and fully-connected shifts) are referred to as DeepShift models. With such DeepShift models that can be implemented with no multiplications, the authors have obtained accuracies of up to 93.6 Top-1/Top-5 accuracies of 70.9 is made on various well-known CNN architectures after converting all their convolution layers and fully connected layers to their bitwise shift counterparts, and we show that in some architectures, the Top-1 accuracy drops by less than 4 have been conducted on PyTorch framework and the code for training and running is submitted along with the paper and will be made available online.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset