LAMBERT: Layout-Aware language Modeling using BERT for information extraction

02/19/2020
by   Łukasz Garncarek, et al.
0

In this paper we introduce a novel approach to the problem of understanding documents where the local semantics is influenced by non-trivial layout. Namely, we modify the Transformer architecture in a way that allows it to use the graphical features defined by the layout, without the need to re-learn the language semantics from scratch, thanks to starting the training process from a model pretrained on classical language modeling tasks.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset