We explore the impact of parameter sparsity on the scaling behavior of
T...
DST methods achieve state-of-the-art results in sparse neural network
tr...
This paper introduces JaxPruner, an open-source JAX-based pruning and sp...
In this work we identify the dormant neuron phenomenon in deep reinforce...
Sparsity has become one of the promising methods to compress and acceler...
The use of sparse neural networks has seen rapid growth in recent years,...
The architecture and the parameters of neural networks are often optimiz...
Transfer-learning methods aim to improve performance in a data-scarce ta...
Meta and transfer learning are two successful families of approaches to
...
Sparse Neural Networks (NNs) can match the generalization of dense NNs u...
Current methods for training recurrent neural networks are based on
back...
Sparse neural networks have been shown to be more parameter and compute
...
This paper explores the task Natural Language Understanding (NLU) by loo...
We investigate the difficulties of training sparse neural networks and m...
Deep Neural Networks are highly over-parameterized and the size of the n...