TY - JOUR
T1 - Explain it as simple as possible, but no simpler – Explanation via model simplification for addressing inferential gap
AU - Sreedharan, Sarath
AU - Srivastava, Siddharth
AU - Kambhampati, Subbarao
N1 - Publisher Copyright:
© 2024
PY - 2025/3
Y1 - 2025/3
N2 - One of the core challenges of explaining decisions made by modern AI systems is the need to address the potential gap in the inferential capabilities of the system generating the decision and the user trying to make sense of it. This inferential capability gap becomes even more critical when it comes to explaining sequential decisions. While there have been some isolated efforts at developing explanation methods suited for complex decision-making settings, most of these current efforts are limited in scope. In this paper, we introduce a general framework for generating explanations in the presence of inferential capability gaps. A framework that is grounded in the generation of simplified representations of the agent model through the application of a sequence of model simplifying transformations. This framework not only allows us to develop an extremely general explanation generation algorithm, but we see that many of the existing works in this direction could be seen as specific instantiations of our more general method. While the ideas presented in this paper are general enough to be applied to any decision-making framework, we will focus on instantiating the framework in the context of stochastic planning problems. As a part of this instantiation, we will also provide an exhaustive characterization of explanatory queries and an analysis of various classes of applicable transformations. We will evaluate the effectiveness of transformation-based explanations through both synthetic experiments and user studies.
AB - One of the core challenges of explaining decisions made by modern AI systems is the need to address the potential gap in the inferential capabilities of the system generating the decision and the user trying to make sense of it. This inferential capability gap becomes even more critical when it comes to explaining sequential decisions. While there have been some isolated efforts at developing explanation methods suited for complex decision-making settings, most of these current efforts are limited in scope. In this paper, we introduce a general framework for generating explanations in the presence of inferential capability gaps. A framework that is grounded in the generation of simplified representations of the agent model through the application of a sequence of model simplifying transformations. This framework not only allows us to develop an extremely general explanation generation algorithm, but we see that many of the existing works in this direction could be seen as specific instantiations of our more general method. While the ideas presented in this paper are general enough to be applied to any decision-making framework, we will focus on instantiating the framework in the context of stochastic planning problems. As a part of this instantiation, we will also provide an exhaustive characterization of explanatory queries and an analysis of various classes of applicable transformations. We will evaluate the effectiveness of transformation-based explanations through both synthetic experiments and user studies.
KW - Abstractions
KW - Contrastive explanations
KW - Explanations for plans
UR - http://www.scopus.com/inward/record.url?scp=85214348650&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85214348650&partnerID=8YFLogxK
U2 - 10.1016/j.artint.2024.104279
DO - 10.1016/j.artint.2024.104279
M3 - Article
AN - SCOPUS:85214348650
SN - 0004-3702
VL - 340
JO - Artificial Intelligence
JF - Artificial Intelligence
M1 - 104279
ER -