A Transparent Pipeline for Identifying Sexism in Social Media: Combining Explainability with Model Prediction

Hadi Mohammadi, Anastasia Giachanou*, Ayoub Bagheri

*Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Featured Application: We show illustrative examples of sexist language to describe the taxonomy and explainability analysis. In this study, we present a new approach that combines multiple Bidirectional Encoder Representations from Transformers (BERT) architectures with a Convolutional Neural Network (CNN) framework designed for sexism detection in text at a granular level. Our method relies on the analysis and identification of the most important terms contributing to sexist content using Shapley Additive Explanations (SHAP) values. This approach involves defining a range of Sexism Scores based on both model predictions and explainability, moving beyond binary classification to provide a deeper understanding of the sexism-detection process. Additionally, it enables us to identify specific parts of a sentence and their respective contributions to this range, which can be valuable for decision makers and future research. In conclusion, this study introduces an innovative method for enhancing the clarity of large language models (LLMs), which is particularly relevant in sensitive domains such as sexism detection. The incorporation of explainability into the model represents a significant advancement in this field. The objective of our study is to bridge the gap between advanced technology and human comprehension by providing a framework for creating AI models that are both efficient and transparent. This approach could serve as a pipeline for future studies to incorporate explainability into language models.

Original languageEnglish
Article number8620
Number of pages27
JournalApplied Sciences (Switzerland)
Volume14
Issue number19
DOIs
Publication statusPublished - Oct 2024

Bibliographical note

Publisher Copyright:
© 2024 by the authors.

Keywords

  • ensemble model
  • explainable AI (XAI)
  • large language models (LLMs)
  • natural language processing (NLP)
  • sexism detection
  • Shapley values

Fingerprint

Dive into the research topics of 'A Transparent Pipeline for Identifying Sexism in Social Media: Combining Explainability with Model Prediction'. Together they form a unique fingerprint.

Cite this