Dichotomize and Generalize: PAC-Bayesian Binary Activated Deep Neural Networks

05/24/2019
by   Gaël Letarte, et al.
2

We present a comprehensive study of multilayer neural networks with binary activation, relying on the PAC-Bayesian theory. Our contributions are twofold: (i) we develop an end-to-end framework to train a binary activated deep neural network, overcoming the fact that binary activation function is non-differentiable; (ii) we provide nonvacuous PAC-Bayesian generalization bounds for binary activated deep neural networks. Noteworthy, our results are obtained by minimizing the expected loss of an architecture-dependent aggregation of binary activated deep neural networks. The performance of our approach is assessed on a thorough numerical experiments protocol on real-life datasets.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset