How to Dissect a Muppet: The Structure of Transformer Embedding Spaces

Timothee Mickus, Denis Paperno, Mathieu Constant

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Pretrained embeddings based on the Transformer architecture have taken the NLP community by storm. We show that they can mathematically be reframed as a sum of vector factors and showcase how to use this reframing to study the impact of each component. We provide evidence that multi-head attentions and feed-forwards are not equally useful in all downstream applications, as well as a quantitative overview of the effects of finetuning on the overall embedding space. This approach allows us to draw connections to a wide range of previous studies, from vector space anisotropy to attention weights.
Original languageEnglish
Pages (from-to)981–996
JournalTransactions of the Association for Computational Linguistics
Volume10
DOIs
Publication statusPublished - 7 Sept 2022

Fingerprint

Dive into the research topics of 'How to Dissect a Muppet: The Structure of Transformer Embedding Spaces'. Together they form a unique fingerprint.

Cite this