LANISTR: Multimodal Learning from Structured and Unstructured Data

05/26/2023
by   Sayna Ebrahimi, et al.
0

Multimodal large-scale pretraining has shown impressive performance gains for unstructured data including language, image, audio, and video. Yet, the scenario most prominent in real-world applications is the existence of combination of structured (including tabular and time-series) and unstructured data, and this has so far been understudied. Towards this end, we propose LANISTR, a novel attention-based framework to learn from LANguage, Image, and STRuctured data. We introduce a new multimodal fusion module with a similarity-based multimodal masking loss that enables LANISTR to learn cross-modal relations from large-scale multimodal data with missing modalities during training and test time. On two publicly available challenging datasets, MIMIC-IV and Amazon Product Review, LANISTR achieves absolute improvements of 6.47 state-of-the-art multimodal models while showing superior generalization capabilities.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset