Leverage Points In Modality Shifts: Comparing Language-Only and Multimodal Word Representations

Aleksey Tikhonov, Denis Paperno, Lisa Bylinina

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

Multimodal embeddings aim to enrich the semantic information in neural representations of language compared to text-only models. While different embeddings exhibit different applicability and performance on downstream tasks, little is known about the systematic representation differences attributed to the visual modality. Our paper compares word embeddings from three vision-and-language models (CLIP, OpenCLIP and Multilingual CLIP, Radford et al. 2021; Ilharco et al. 2021; Carlsson et al. 2022) and three text-only models, with static (FastText, Bojanowski et al. 2017) as well as contextual representations (multilingual BERT Devlin et al. 2018; XLM-RoBERTa, Conneau et al. 2019). This is the first large-scale study of the effect of visual grounding on language representations, including 46 semantic parameters. We identify meaning properties and relations that characterize words whose embeddings are most affected by the inclusion of visual modality in the training data; that is, points where visual grounding turns out most important. We find that the effect of visual modality correlates most with denotational semantic properties related to concreteness, but is also detected for several specific semantic classes, as well as for valence, a sentiment-related connotational property of linguistic expressions.
Original languageEnglish
Title of host publicationProceedings of the 12th Joint Conference on Lexical and Computational Semantics (*SEM 2023)
PublisherAssociation for Computational Linguistics
Pages11–17
Number of pages7
DOIs
Publication statusPublished - Jul 2023

Fingerprint

Dive into the research topics of 'Leverage Points In Modality Shifts: Comparing Language-Only and Multimodal Word Representations'. Together they form a unique fingerprint.

Cite this