Artificial General Intelligence (AGI) requires comprehensive understandi...
The convergence of text, visual, and audio data is a key step towards
hu...
A common thread of retrieval-augmented methods in the existing literatur...
Human intelligence is multimodal; we integrate visual, linguistic, and
a...
Commonsense reasoning (CSR) requires the model to be equipped with gener...
Pre-trained language models (PLMs) aim to learn universal language
repre...
Current Open-Domain Question Answering (ODQA) model paradigm often conta...
It is often observed in knowledge-centric tasks (e.g., common sense ques...
Recent advances in large-scale pre-training such as GPT-3 allow seemingl...
Multimodal pre-training has propelled great advancement in
vision-and-la...
In this paper, we propose Cross-Thought, a novel approach to pre-trainin...
Existing language model compression methods mostly use a simple L2 loss ...
Transformer has become ubiquitous in the deep learning field. One of the...
Existing approaches to real-time question answering (RTQA) rely on learn...
Large-scale cross-lingual language models (LM), such as mBERT, Unicoder ...
In this paper, we present Hierarchical Graph Network (HGN) for multi-hop...
This paper presents an extension of the Stochastic Answer Network (SAN),...