Abstract
Featured Application: We show illustrative examples of sexist language to describe the taxonomy and explainability analysis. In this study, we present a new approach that combines multiple Bidirectional Encoder Representations from Transformers (BERT) architectures with a Convolutional Neural Network (CNN) framework designed for sexism detection in text at a granular level. Our method relies on the analysis and identification of the most important terms contributing to sexist content using Shapley Additive Explanations (SHAP) values. This approach involves defining a range of Sexism Scores based on both model predictions and explainability, moving beyond binary classification to provide a deeper understanding of the sexism-detection process. Additionally, it enables us to identify specific parts of a sentence and their respective contributions to this range, which can be valuable for decision makers and future research. In conclusion, this study introduces an innovative method for enhancing the clarity of large language models (LLMs), which is particularly relevant in sensitive domains such as sexism detection. The incorporation of explainability into the model represents a significant advancement in this field. The objective of our study is to bridge the gap between advanced technology and human comprehension by providing a framework for creating AI models that are both efficient and transparent. This approach could serve as a pipeline for future studies to incorporate explainability into language models.
Original language | English |
---|---|
Article number | 8620 |
Number of pages | 27 |
Journal | Applied Sciences (Switzerland) |
Volume | 14 |
Issue number | 19 |
DOIs | |
Publication status | Published - Oct 2024 |
Bibliographical note
Publisher Copyright:© 2024 by the authors.
Keywords
- ensemble model
- explainable AI (XAI)
- large language models (LLMs)
- natural language processing (NLP)
- sexism detection
- Shapley values