Representation Learning for Type-Driven Composition

Gijs Wijnholds, Mehrnoosh Sadrzadeh, Stephen Clark

Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

Abstract

This paper is about learning word representations using grammatical type information. We use the syntactic types of Combinatory Categorial Grammar to develop multilinear representations, i.e. maps with n arguments, for words with different functional types. The multilinear maps of words compose with each other to form sentence representations. We extend the skipgram algorithm from vectors to multi- linear maps to learn these representations and instantiate it on unary and binary maps for transitive verbs. These are evaluated on verb and sentence similarity and disambiguation tasks and a subset of the SICK relatedness dataset. Our model performs better than previous type- driven models and is competitive with state of the art representation learning methods such as BERT and neural sentence encoders.
Original languageEnglish
Title of host publicationProceedings of the 24th Conference on Computational Natural Language Learning
EditorsRaquel Fernández, Tal Linzen
PublisherAssociation for Computational Linguistics
Pages313–324
Number of pages12
DOIs
Publication statusPublished - Nov 2020

Fingerprint

Dive into the research topics of 'Representation Learning for Type-Driven Composition'. Together they form a unique fingerprint.

Cite this