The Explabox: Model-Agnostic Machine Learning Transparency & Analysis

Research output: Working paperPreprintAcademic

Abstract

We present the Explabox: an open-source toolkit for transparent and responsible machine learning (ML) model development and usage. Explabox aids in achieving explainable, fair and robust models by employing a four-step strategy: explore, examine, explain and expose. These steps offer model-agnostic analyses that transform complex 'ingestibles' (models and data) into interpretable 'digestibles'. The toolkit encompasses digestibles for descriptive statistics, performance metrics, model behavior explanations (local and global), and robustness, security, and fairness assessments. Implemented in Python, Explabox supports multiple interaction modes and builds on open-source packages. It empowers model developers and testers to operationalize explainability, fairness, auditability, and security. The initial release focuses on text data and models, with plans for expansion. Explabox's code and documentation are available open-source at https://explabox.readthedocs.io/
Original languageEnglish
PublisherarXiv
Number of pages5
DOIs
Publication statusPublished - 2024

Keywords

  • explainable AI (XAI)
  • interpretability
  • fairness
  • robustness
  • AI safety
  • auditability

Fingerprint

Dive into the research topics of 'The Explabox: Model-Agnostic Machine Learning Transparency & Analysis'. Together they form a unique fingerprint.

Cite this