TextFocus: Assessing the Faithfulness of Feature Attribution Methods Explanations in Natural Language Processing

Ettore Mariotti*, Anna Arias-Duart, Michele Cafagna, Albert Gatt, Dario Garcia-Gasulla, Jose Maria Alonso-Moral

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Among the existing eXplainable AI (XAI) approaches, Feature Attribution methods are a popular option due to their interpretable nature. However, each method leads to a different solution, thus introducing uncertainty regarding their reliability and coherence with respect to the underlying model. This work introduces TextFocus, a metric for evaluating the faithfulness of Feature Attribution methods for Natural Language Processing (NLP) tasks involving classification. To address the absence of ground truth explanations for such methods, we introduce the concept of textual mosaics. A mosaic is composed of a combination of sentences belonging to different classes, which provides an implicit ground truth for attribution. The accuracy of explanations can be then evaluated by comparing feature attribution scores with the known class labels in the mosaic. The performance of six feature attribution methods is systematically compared on three sentence classification tasks by using TextFocus, with Integrated Gradients being the best overall method in terms of faithfulness and computational requirements. The proposed methodology fills a gap in NLP evaluation, by providing an objective way to assess Feature Attribution methods while finding their optimal parameters.

Original languageEnglish
Pages (from-to)138870-138880
Number of pages11
JournalIEEE Access
Volume12
DOIs
Publication statusPublished - 31 May 2024

Keywords

  • Artificial intelligence (AI)
  • explainable AI (XAI)
  • explanation faithfulness
  • feature attribution
  • feature importance
  • natural language processing (NLP)
  • trustworthy AI

Fingerprint

Dive into the research topics of 'TextFocus: Assessing the Faithfulness of Feature Attribution Methods Explanations in Natural Language Processing'. Together they form a unique fingerprint.

Cite this