Baby Llama: knowledge distillation from an ensemble of teachers trained on a small dataset with no performance penalty

08/03/2023
by   Inar Timiryasov, et al.
0

We present our proposed solution to the BabyLM challenge [arXiv:2301.11796], whose goal was to improve the sample efficiency of language models. We trained an ensemble consisting of a GPT-2 and small LLaMA models on the developmentally-plausible, 10M-word BabyLM dataset, then distilled it into a small, 58M-parameter LLaMA model, which exceeds in performance both of its teachers as well as a similar model trained without distillation. This suggests that distillation can not only retain the full performance of the teacher model when the latter is trained on a sufficiently small dataset; it can exceed it, and lead to significantly better performance than direct training.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset