In this work, we develop and release Llama 2, a collection of pretrained...
In-context learning (ICL) improves language models' performance on a var...
Recent work has shown that fine-tuning large pre-trained language models...
Self-supervised pretraining has made few-shot learning possible for many...
Large language models, which are often trained for hundreds of thousands...
Mixture of Experts layers (MoEs) enable efficient scaling of language mo...
Large-scale autoregressive language models such as GPT-3 are few-shot
le...
We propose EXAMS – a new benchmark dataset for cross-lingual and
multili...
We describe our system for finding good answers in a community forum, as...
There are different definitions of what a troll is. Certainly, a troll c...
In this work, we propose to use linguistic annotations as a basis for a
...
We present a new kind of question answering dataset, OpenBookQA, modeled...
We introduce a neural reading comprehension model that integrates extern...
Reading comprehension is a challenging task in natural language processi...
We transfer a key idea from the field of sentiment analysis to a new dom...
This paper describes two supervised baseline systems for the Story Cloze...