On the tradeoff between correctness and completeness in argumentative explainable AI
File(s)8151.pdf (1.2 MB)
Published version
Author(s)
Potyka, N
Yin, X
Toni, F
Type
Conference Paper
Abstract
Explainable AI aims at making the decisions of autonomous systems human-understandable. Argumentation frameworks are a natural tool for this purpose. Among them, bipolar abstract argumentation frameworks seem well suited to explain the effect of features on a classification decision and their formal properties can potentially be used to derive formal guarantees for explanations. Two particular interesting properties are correctness (if the explanation says that X affects Y, then X affects Y ) and completeness (if X affects Y, then the explanation says that X affects Y ). The reinforcement property of bipolar argumentation frameworks has been used as a natural correctness counterpart in previous work. Applied to the classification context, it basically states that attacking features should decrease and supporting features should increase the confidence of a classifier. In this short discussion paper, we revisit this idea, discuss potential limitations when considering reinforcement without a corresponding completeness property and how these limitations can potentially be overcome.
Date Issued
2022-09-12
Date Acceptance
2022-09-01
Citation
CEUR Workshop Proceedings, 2022, 3209, pp.1-8
ISSN
1613-0073
Publisher
CEUR Workshop Proceedings
Start Page
1
End Page
8
Journal / Book Title
CEUR Workshop Proceedings
Volume
3209
Copyright Statement
© 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
License URL
Identifier
https://ceur-ws.org/Vol-3209/8151.pdf
Source
1st International Workshop on Argumentation for eXplainable AI
Publication Status
Published
Start Date
2022-09-12
Finish Date
2022-09-12
Coverage Spatial
Cardiff, United Kingdom
Date Publish Online
2022-09-12