Muddling Label Regularization: Deep Learning for Tabular Datasets

06/08/2021
by   Karim Lounici, et al.
0

Deep Learning (DL) is considered the state-of-the-art in computer vision, speech recognition and natural language processing. Until recently, it was also widely accepted that DL is irrelevant for learning tasks on tabular data, especially in the small sample regime where ensemble methods are acknowledged as the gold standard. We present a new end-to-end differentiable method to train a standard FFNN. Our method, Muddling labels for Regularization (), penalizes memorization through the generation of uninformative labels and the application of a differentiable close-form regularization scheme on the last hidden layer during training. outperforms classical NN and the gold standard (GBDT, RF) for regression and classification tasks on several datasets from the UCI database and Kaggle covering a large range of sample sizes and feature to sample ratios. Researchers and practitioners can use on its own as an off-the-shelf solution or integrate it into the most advanced ML pipelines.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset