Estimating Causal Responsibility for Explaining Autonomous Behavior

Saaduddin Mahmud*, Samer B. Nashed, Claudia V. Goldman, Shlomo Zilberstein

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

There has been growing interest in causal explanations of stochastic, sequential decision-making systems. Structural causal models and causal reasoning offer several theoretical benefits when exact inference can be applied. Furthermore, users overwhelmingly prefer the resulting causal explanations over other state-of-the-art systems. In this work, we focus on one such method, MeanRESP, and its approximate versions that drastically reduce compute load and assign a responsibility score to each variable, which helps identify smaller sets of causes to be used as explanations. However, this method, and its approximate versions in particular, lack deeper theoretical analysis and broader empirical tests. To address these shortcomings, we provide three primary contributions. First, we offer several theoretical insights on the sample complexity and error rate of approximate MeanRESP. Second, we discuss several automated metrics for comparing explanations generated from approximate methods to those generated via exact methods. While we recognize the significance of user studies as the gold standard for evaluating explanations, our aim is to leverage the proposed metrics to systematically compare explanation-generation methods along important quantitative dimensions. Finally, we provide a more detailed discussion of MeanRESP and how its output under different definitions of responsibility compares to existing widely adopted methods that use Shapley values.

Original languageEnglish
Title of host publicationExplainable and Transparent AI and Multi-Agent Systems - 5th International Workshop, EXTRAAMAS 2023, Revised Selected Papers
EditorsDavide Calvaresi, Amro Najjar, Andrea Omicini, Rachele Carli, Giovanni Ciatto, Reyhan Aydogan, Yazan Mualla, Kary Främling
PublisherSpringer Science and Business Media Deutschland GmbH
Pages78-94
Number of pages17
ISBN (Print)9783031408779
DOIs
StatePublished - 2023
Externally publishedYes
EventProceedings of the 5th International Workshop on EXTRAAMAS 2023 - London, United Kingdom
Duration: 29 May 202329 May 2023

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14127 LNAI
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

ConferenceProceedings of the 5th International Workshop on EXTRAAMAS 2023
Country/TerritoryUnited Kingdom
CityLondon
Period29/05/2329/05/23

Bibliographical note

Publisher Copyright:
© 2023, The Author(s), under exclusive license to Springer Nature Switzerland AG.

Keywords

  • Causal Inference
  • Explainable AI
  • MDPs

Fingerprint

Dive into the research topics of 'Estimating Causal Responsibility for Explaining Autonomous Behavior'. Together they form a unique fingerprint.

Cite this