Abstract
We propose an approach to faithfully explaining text classification models, using a specifically designed neural network to
find explanations in the form of machine-annotated rationales
during the prediction process. This results in faithful explanations that are similar to human-annotated rationales, while not
requiring human explanation examples during training. The
quality of found explanations is measured on faithfulness,
quantitative similarity to human explanations, and through a
user evaluation.
find explanations in the form of machine-annotated rationales
during the prediction process. This results in faithful explanations that are similar to human-annotated rationales, while not
requiring human explanation examples during training. The
quality of found explanations is measured on faithfulness,
quantitative similarity to human explanations, and through a
user evaluation.
| Original language | English |
|---|---|
| Number of pages | 8 |
| Publication status | Published - 2021 |
| Event | 35th AAAI Conference on Artificial Intelligence - Duration: 8 Feb 2021 → 9 Feb 2021 |
Conference
| Conference | 35th AAAI Conference on Artificial Intelligence |
|---|---|
| Period | 8/02/21 → 9/02/21 |
Fingerprint
Dive into the research topics of 'Machine-annotated Rationales: Faithfully Explaining Text Classification'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver