TY - JOUR
T1 - From “no clear winner” to an effective Explainable Artificial Intelligence process
T2 - An empirical journey
AU - Dodge, Jonathan
AU - Anderson, Andrew
AU - Khanna, Roli
AU - Irvine, Jed
AU - Dikkala, Rupika
AU - Lam, Kin Ho
AU - Tabatabai, Delyar
AU - Ruangrotsakun, Anita
AU - Shureih, Zeyad
AU - Kahng, Minsuk
AU - Fern, Alan
AU - Burnett, Margaret
N1 - Publisher Copyright:
© 2021 The Authors. Applied AI Letters published by John Wiley & Sons Ltd.
PY - 2021/12
Y1 - 2021/12
N2 - “In what circumstances would you want this AI to make decisions on your behalf?” We have been investigating how to enable a user of an Artificial Intelligence-powered system to answer questions like this through a series of empirical studies, a group of which we summarize here. We began the series by (a) comparing four explanation configurations of saliency explanations and/or reward explanations. From this study we learned that, although some configurations had significant strengths, no one configuration was a clear “winner.” This result led us to hypothesize that one reason for the low success rates Explainable AI (XAI) research has in enabling users to create a coherent mental model is that the AI itself does not have a coherent model. This hypothesis led us to (b) build a model-based agent, to compare explaining it with explaining a model-free agent. Our results were encouraging, but we then realized that participants' cognitive energy was being sapped by having to create not only a mental model, but also a process by which to create that mental model. This realization led us to (c) create such a process (which we term After-Action Review for AI or “AAR/AI”) for them, integrate it into the explanation environment, and compare participants' success with AAR/AI scaffolding vs without it. Our AAR/AI studies' results showed that AAR/AI participants were more effective assessing the AI than non-AAR/AI participants, with significantly better precision and significantly better recall at finding the AI's reasoning flaws.
AB - “In what circumstances would you want this AI to make decisions on your behalf?” We have been investigating how to enable a user of an Artificial Intelligence-powered system to answer questions like this through a series of empirical studies, a group of which we summarize here. We began the series by (a) comparing four explanation configurations of saliency explanations and/or reward explanations. From this study we learned that, although some configurations had significant strengths, no one configuration was a clear “winner.” This result led us to hypothesize that one reason for the low success rates Explainable AI (XAI) research has in enabling users to create a coherent mental model is that the AI itself does not have a coherent model. This hypothesis led us to (b) build a model-based agent, to compare explaining it with explaining a model-free agent. Our results were encouraging, but we then realized that participants' cognitive energy was being sapped by having to create not only a mental model, but also a process by which to create that mental model. This realization led us to (c) create such a process (which we term After-Action Review for AI or “AAR/AI”) for them, integrate it into the explanation environment, and compare participants' success with AAR/AI scaffolding vs without it. Our AAR/AI studies' results showed that AAR/AI participants were more effective assessing the AI than non-AAR/AI participants, with significantly better precision and significantly better recall at finding the AI's reasoning flaws.
UR - http://www.scopus.com/inward/record.url?scp=85125687843&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85125687843&partnerID=8YFLogxK
U2 - 10.1002/ail2.36
DO - 10.1002/ail2.36
M3 - Letter
AN - SCOPUS:85125687843
SN - 2689-5595
VL - 2
JO - Applied AI Letters
JF - Applied AI Letters
IS - 4
M1 - e36
ER -