Argumentative explanations for pattern-based text classifiers

Recent works in Explainable AI mostly address the transparency issue of black-box models or create explanations for any kind of models (i.e., they are model-agnostic), while leaving explanations of interpretable models largely underexplored. In this paper, we fill this gap by focusing on expla...

Full description

Bibliographic Details
Main Authors: Piyawat Lertvittayakumjorn, Francesca Toni
Format: Article
Language:English
Published: IOS Press 2023-01-01
Series:Argument & Computation
Online Access:https://doi.org/10.3233/AAC-220004
_version_ 1797860860499066880
author Piyawat Lertvittayakumjorn
Francesca Toni
author_facet Piyawat Lertvittayakumjorn
Francesca Toni
author_sort Piyawat Lertvittayakumjorn
collection DOAJ
description Recent works in Explainable AI mostly address the transparency issue of black-box models or create explanations for any kind of models (i.e., they are model-agnostic), while leaving explanations of interpretable models largely underexplored. In this paper, we fill this gap by focusing on explanations for a specific interpretable model, namely pattern-based logistic regression (PLR) for binary text classification. We do so because, albeit interpretable, PLR is challenging when it comes to explanations. In particular, we found that a standard way to extract explanations from this model does not consider relations among the features, making the explanations hardly plausible to humans. Hence, we propose AXPLR, a novel explanation method using (forms of) computational argumentation to generate explanations (for outputs computed by PLR) which unearth model agreements and disagreements among the features. Specifically, we use computational argumentation as follows: we see features (patterns) in PLR as arguments in a form of quantified bipolar argumentation frameworks (QBAFs) and extract attacks and supports between arguments based on specificity of the arguments; we understand logistic regression as a gradual semantics for these QBAFs, used to determine the arguments’ dialectic strength; and we study standard properties of gradual semantics for QBAFs in the context of our argumentative re-interpretation of PLR, sanctioning its suitability for explanatory purposes. We then show how to extract intuitive explanations (for outputs computed by PLR) from the constructed QBAFs. Finally, we conduct an empirical evaluation and two experiments in the context of human-AI collaboration to demonstrate the advantages of our resulting AXPLR method.
first_indexed 2024-04-09T21:53:47Z
format Article
id doaj.art-1221f831233c49d28451adb064d5533b
institution Directory Open Access Journal
issn 1946-2166
1946-2174
language English
last_indexed 2024-04-09T21:53:47Z
publishDate 2023-01-01
publisher IOS Press
record_format Article
series Argument & Computation
spelling doaj.art-1221f831233c49d28451adb064d5533b2023-03-24T09:48:03ZengIOS PressArgument & Computation1946-21661946-21742023-01-0117210.3233/AAC-220004Argumentative explanations for pattern-based text classifiersPiyawat Lertvittayakumjorn0Francesca Toni1Department of Computing, Imperial College London, United KingdomDepartment of Computing, Imperial College London, United Kingdom Recent works in Explainable AI mostly address the transparency issue of black-box models or create explanations for any kind of models (i.e., they are model-agnostic), while leaving explanations of interpretable models largely underexplored. In this paper, we fill this gap by focusing on explanations for a specific interpretable model, namely pattern-based logistic regression (PLR) for binary text classification. We do so because, albeit interpretable, PLR is challenging when it comes to explanations. In particular, we found that a standard way to extract explanations from this model does not consider relations among the features, making the explanations hardly plausible to humans. Hence, we propose AXPLR, a novel explanation method using (forms of) computational argumentation to generate explanations (for outputs computed by PLR) which unearth model agreements and disagreements among the features. Specifically, we use computational argumentation as follows: we see features (patterns) in PLR as arguments in a form of quantified bipolar argumentation frameworks (QBAFs) and extract attacks and supports between arguments based on specificity of the arguments; we understand logistic regression as a gradual semantics for these QBAFs, used to determine the arguments’ dialectic strength; and we study standard properties of gradual semantics for QBAFs in the context of our argumentative re-interpretation of PLR, sanctioning its suitability for explanatory purposes. We then show how to extract intuitive explanations (for outputs computed by PLR) from the constructed QBAFs. Finally, we conduct an empirical evaluation and two experiments in the context of human-AI collaboration to demonstrate the advantages of our resulting AXPLR method.https://doi.org/10.3233/AAC-220004
spellingShingle Piyawat Lertvittayakumjorn
Francesca Toni
Argumentative explanations for pattern-based text classifiers
Argument & Computation
title Argumentative explanations for pattern-based text classifiers
title_full Argumentative explanations for pattern-based text classifiers
title_fullStr Argumentative explanations for pattern-based text classifiers
title_full_unstemmed Argumentative explanations for pattern-based text classifiers
title_short Argumentative explanations for pattern-based text classifiers
title_sort argumentative explanations for pattern based text classifiers
url https://doi.org/10.3233/AAC-220004
work_keys_str_mv AT piyawatlertvittayakumjorn argumentativeexplanationsforpatternbasedtextclassifiers
AT francescatoni argumentativeexplanationsforpatternbasedtextclassifiers