Abstract
This paper provides evidence indicating that the most commonly used metric for validating feature attribution methods in eXplainable AI (XAI) is misleading when applied to time series data. To evaluate whether an XAI method attributes importance to relevant features, these are systematically perturbed while measuring the impact on the performance of the classifier. The assumption is that a drastic performance reduction with increasing perturbation of relevant features indicates that these are indeed relevant. We demonstrate empirically that this assumption is incomplete without considering low relevance features in the used metrics. We introduce a novel metric, the Perturbation Effect Size, and demonstrate how it complements existing metrics to offer a more faithful assessment of importance attribution. Finally, we contribute a comprehensive evaluation of attribution methods on time series data, considering the influence of perturbation methods and region size selection.
Original language | English |
---|---|
Title of host publication | CIKM '22: Proceedings of the 31st ACM International Conference on Information & Knowledge Management |
Place of Publication | New York, NY |
Publisher | Association of Computing Machinery |
Pages | 1798-1807 |
ISBN (Electronic) | 978-1-4503-9236-5 |
DOIs | |
Publication status | Published - 17 Oct 2022 |
Event | 31st ACM International Conference on Information and Knowledge Management: CIKM 2022 - Atlanta, United States Duration: 17 Oct 2022 → 21 Oct 2022 |
Conference
Conference | 31st ACM International Conference on Information and Knowledge Management |
---|---|
Abbreviated title | CIKM '22 |
Country/Territory | United States |
City | Atlanta |
Period | 17/10/22 → 21/10/22 |
Keywords
- deep learning
- explainable AI
- trustworthy AI
- feature attribution