Abstract
Existing datasets for scoring text pairs in terms of semantic similarity contain instances whose resolution differs according to the degree of difficulty. This paper proposes to distinguish obvious from non-obvious text pairs based on superficial lexical overlap and ground-truth labels. We characterise existing datasets in terms of containing difficult cases and find that recently proposed models struggle to capture the non-obvious cases of semantic similarity. We describe metrics that emphasise cases of similarity which require more complex inference and propose that these are used for evaluating systems for semantic similarity.
Original language | English |
---|---|
Title of host publication | Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics |
Place of Publication | Florence, Italy |
Publisher | Association for Computational Linguistics |
Pages | 2792-2798 |
Number of pages | 7 |
DOIs | |
Publication status | Published - 28 Jul 2019 |