Ensuring the interpretability of trained machine learning models is often paramount, particularly in high-stakes domains such as counter-terrorism and other forms of law enforcement. Post hoc techniques have emerged as a promising avenue for justifying the predictions of complex models. However, while these approaches provide valuable insights, they often lack the ability to directly reference familiar domain rules and make use of these rules to guide explanations. This paper introduces a method for incorporating arguments about the applicability of domain rules in justifying classifier predictions
Original language | English |
---|
Title of host publication | CEUR Workshop Proceedings |
---|
Publisher | CEUR WS |
---|
Pages | 90-99 |
---|
Volume | 3769 |
---|
Publication status | Published - 2024 |
---|
Name | CEUR Workshop Proceedings |
---|
Publisher | CEUR |
---|
Volume | 3769 |
---|
ISSN (Electronic) | 1613-0073 |
---|