Principles of Explanation in Human-AI Systems
2021, arXiv (Cornell University)
Abstract
Explainable Artificial Intelligence (XAI) has re-emerged in response to the development of modern AI and ML systems. These systems are complex and sometimes biased, but they nevertheless make decisions that impact our lives. XAI systems are frequently algorithm-focused; starting and ending with an algorithm that implements a basic untested idea about explainability. These systems are often not tested to determine whether the algorithm helps users accomplish any goals, and so their explainability remains unproven. We propose an alternative: to start with human-focused principles for the design, testing, and implementation of XAI systems, and implement algorithms to serve that purpose. In this paper, we review some of the basic concepts that have been used for user-centered XAI systems over the past 40 years of research. Based on these, we describe the "Self-Explanation Scorecard", which can help developers understand how they can empower users to by enabling self-explanation. Finally, we present a set of empiricallygrounded, user-centered design principles that may guide developers to create successful explainable systems. User-Centered Explanation in AI Although usability testing is a cornerstone of user-centered design, evaluation often comes too late to provide guidance about implementing a usable system. In response, researchers and designers have proposed guidelines that codify research on human users and advocate for the involvement of users in system development from the beginning (e.g., Greenbaum and Kyng 1991; Hoffman et al. 2010). The most famous and detailed set of guidelines may be Apple's Human Interface Guidelines (cf. Mountford 1998), but others have proposed simpler principles such as Neilson's (1994) interface design heuristics or Karat's (1998) "User's Bill of Rights". With the advent of new, powerful AI systems that are complex and difficult to understand, the field of Explainable AI (XAI) has re-emerged as an important area of human-machine interaction. Much of the interest in XAI has focused on deep learning systems. Consequently, most explanations have concentrated on technologies to visualize or otherwise expose deep networks structures, features, or
References (66)
- Adams, B. D., Bruyn, L. E., Houde, S., Angelopoulos, P., Iwasa- Madge, K., and McCann, C. 2003. Trust in automated systems. Canada Ministry of National Defence.
- Alam, L. 2020. Investigating the Impact of Explanation on Re- pairing Trust in Ai Diagnostic Systems for Re-Diagnosis (Publi- cation No. 28088930) [Master's Thesis, Michigan Technological University].
- Aloimonos, J., Weiss, I., and Bandyopadhyay, A. 1988. Active vision. International journal of computer vision, 1(4), 333-356.
- Brézillon, P. 1994. Context needs in cooperative building of ex- planations. In First European Conference on Cognitive Science in Industry (pp. 443-450).
- Clancey, W. J. 1986. From GUIDON to NEOMYCIN and HER- ACLES in twenty short lessons. AI Magazine, 7(3), 40. Clancey, W. J. 2020. Designing agents for people: Case studies of the Brahms work practice simulation framework. [Kindle].
- Chari, S., Seneviratne, O., Gruen, D. M., Foreman, M. A., Das, A. K., & McGuinness, D. L. (2020, November). Explanation On- tology: A Model of Explanations for User-Centered AI. In Inter- national Semantic Web Conference (pp. 228-243). Springer.
- Chi, M. T., Bassok, M., Lewis, M. W., Reimann, P., and Glaser, R. 1989. Self-explanations: How students study and use examples in learning to solve problems. Cognitive Science, 13(2), 145-182.
- Conant, R. C., and Ashby, W R. 1970. Every good regulator of a system must be a model of that system. International Journal of Systems Science, 1(2), 89-97.
- Deal, S.V. and Hoffman, R.R. (2010, September/October). The Practitioner's Cycles Part 3: Implementation problems. IEEE In- telligent Systems, pp. 77-81.
- Doshi-Velez, F., and Kim, B. 2017. A roadmap for a rigorous sci- ence of interpretability. ArXiv Preprint ArXiv:1702.08608. https://arxiv.org/abs/1702.08608
- Doyle, J. K., and Ford, D., 1998, Mental models concepts for sys- tem dynamics research, System Dynamics Review, 14(1), 3-29.
- Doyle, J., Radzicki, M., & Trees, W. 2008. Measuring change in mental models of complex dynamic systems. Complex Decision Making, 269-294.
- Doyle, D., Tsymbal, A., & Cunningham, P. 2003. A review of ex- planation and explanation in case-based reasoning. Trinity Col- lege Dublin, Department of Computer Science Technical Report TCD-CS-2003-41. http://www.tara.tcd.ie/handle/2262/12919
- Fallon, C. K., and Blaha, L. M. 2018. Improving Automation Transparency: Addressing Some of Machine Learning's Unique Challenges. International Conference on Augmented Cognition, 245-254. Springer.
- Forrester, J. 1961. Industrial Dynamics. Cambridge, MA: Pro- ductivity Press
- Findlay, J. M., Findlay, J. M., and Gilchrist, I. D. 2003. Active vi- sion: The psychology of looking and seeing (No. 37). Oxford University Press.
- Goodman, B., & Flaxman, S. 2017. European Union regulations on algorithmic decision-making and a "right to explanation". AI magazine, 38(3), 50-57.
- Greenbaum, J. & Kyng, M. (Eds.). 1991. Design at work: Coop- erative design of computer systems. Hillsdale, NJ: Lawrence Erl- baum Associates.
- Grice. H. P. 1975. Logic and Conversation. In Syntax and seman- tics 3: Speech arts. 41-58.
- Guerlain, S. 1995. Using the critiquing approach to cope with brittle expert systems. In Proceedings of the Human Factors and Ergonomics Society Annual Meeting (Vol. 39, pp. 233-237). SAGE Publications Sage CA: Los Angeles, CA.
- Greenbaum, J. and Kyng, M. (Eds.) 1991. Design at work: Coop- erative design of computer systems. Hillsdale, NJ: Lawrence Erl- baum Associates.
- Hendricks, L. A., Akata, Z., Rohrbach, M., Donahue, J., Schiele, B., and Darrell, T. 2016. Generating visual explanations. Euro- pean Conference on Computer Vision, 3-19. http://link.springer.- com/chapter/10.1007/978-3-319-46493-0_1
- Hoffman, R. R., Deal, S. V., Potter, S., and Roth, E. M. 2010. The practitioner's cycles, part 2: Solving envisioned world problems. IEEE Intelligent Systems, 25(3), 6-11.
- Hoffman, R.R. (ed) 2012. Collected Essays on Human-Centered Computing, 2001-2011. New York: IEEE Computer Soc. Press.
- Hoffman, R.R. 2017. A taxonomy of emergent trusting in the hu- man-machine relationship. In P. Smith and R.R. Hoffman, Eds., Cognitive systems engineering: The future for a changing world (137-164). Boca Raton, FL: Taylor & Francis.
- Hoffman, R. R., Klein, G., and Mueller, S. T. 2018b. Explaining explanation for "Explainable AI". Proceedings of the 2018 con- ference of the Human Factors and Ergonomic Society (HFES), Philadelphia PA, October 2018.
- Hoffman, R. R., Mueller, S. T., and Klein, G. 2017. Explaining Explanation, Part 2: Empirical Foundations. IEEE Intelligent Sys- tems, 32(4), 78-86.
- Hoffman, R., Miller, T., Mueller, S. T., Klein, G. and Clancey, W. J. 2018c. Explaining Explanation, Part 4: A Deep Dive on Deep Nets. IEEE Intelligent Systems, 33(3), 87-95.
- Hoffman, R. R., Mueller, S. T., Klein, G., and Litman, J. 2018a. Metrics for explainable AI: Challenges and prospects. arXiv:1812.04608.
- Hoffman, R. R., Clancey, W. J., and Mueller, S. T. 2020. Ex- plaining AI as an Exploratory Process: The Peircean Abduction Model. arXiv:2009.14795.
- Karat, C. M. 1998. Guaranteeing rights for the user. Communica- tions of the ACM, 41(12), 29-31.
- Karsenty, L., and Brezillon, P. J. 1995. Cooperative problem solving and explanation. Expert Systems with Applications, 8(4), 445-462.
- Kass, R., and Finin, T. 1988. The need for user models in generat- ing expert system explanation. International Journal of Expert Systems, 1(4), 345-375.
- Kaur, H., Nori, H., Jenkins, S., Caruana, R., Wallach, H., & Wortman Vaughan, J. 2020. Interpreting Interpretability: Under- standing Data Scientists' Use of Interpretability Tools for Ma- chine Learning. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems (pp. 1-14).
- Kim, B., Khanna, R., and Koyejo, O. O. 2016. Examples are not enough, learn to criticize! Criticism for interpretability. In Ad- vances in Neural Information Processing Systems (2280-2288).
- Klein, G. and Hoffman, R. R. 2008. Macrocognition, mental models, and cognitive task analysis methodology. In J. M. Schraagen, L. G. Militello, T. Ormerod and R. Lipshitz (Eds.), Naturalistic decision making and macrocognition (pp. 57-80). Aldershot, England: Ashgate.
- Klein, G. Hoffman, R. R., and Mueller. S. T. 2019. Scorecard for self-explaining capabilities of AI Systems. Technical Report Pre- pared by Task Area 2, DARPA XAI Program.
- Kulesza, T., Burnett, M., Wong, W.-K., and Stumpf, S. 2015. Principles of explanatory debugging to personalize interactive machine learning. Proceedings of the 20th International Confer- ence on Intelligent User Interfaces, 126-137.
- Lakkaraju, H., Kamar, E., Caruana, R., and Leskovec, J. 2017. In- terpretable & explorable approximations of black box models. ArXiv Preprint ArXiv:1707.01154.
- Langlotz, C. P., and Shortliffe, E. H. 1989. The critiquing ap- proach to automated advice and explanation: Rationale and exam- ples. In Expert Knowledge an Explanation: The Knowledge-Lan- guage Interface, Charlie Ellis (Ed). Ellis Horwood Limited.
- Leake, D. B. 1995. Abduction, experience, and goals: A model of everyday abductive explanation. Journal of Experimental & The- oretical Artificial Intelligence, 7(4), 407-428.
- Lim, B. Y., Dey, A. K., and Avrahami, D. 2009. Why and why not explanations improve the intelligibility of context-aware intel- ligent systems. In Proceedings of the SIGCHI Conference on Hu- man Factors in Computing Systems (pp. 2119-2128). ACM. Meyers, C., and Jones, T. B. 1993. Promoting Active Learning. Strategies for the College Classroom. Jossey-Bass Inc., Publish- ers, 350 Sansome Street, San Francisco, CA 94104.
- Miller, T. 2017. Explanation in artificial intelligence: Insights from the social sciences. ArXiv:1706.07269 [Cs]. http://arxiv.org/ abs/1706.07269
- Moore, J. D., and Swartout, W. R. 1988. Explanation in expert systems: A survey. University of Southern California Marina del rey Information Sciences Institute. DTIC #ADA206283.
- Mountford, S. J. 1998. A history of the Apple human interface group. ACM SIGCHI Bulletin, 30(2), 144-146.
- Mueller, S. T., Agarwal, P. Linja, A, Dave, N., and Alam, L. 2020. The unreasonable ineptitude of deep image classification networks. Proceedings of the 64 th annual meeting of the Human Factors Society.
- Mueller, S.T., Hoffman, R.R., Clancey, W, Emrey, A., and Klein, G. 2019. "Explanation in Human-AI Systems: A Literature Meta- Review, Synopsis of Key Ideas and Publications, and Bibliogra- phy for Explainable AI." Report on Award No. FA8650-17-2- 7711, DARPA XAI Program. DTIC AD1073994.
- Mueller, S.T. and Klein, G. 2011. Improving Users' Mental Mod- els of Intelligent Software Tools," Intelligent Systems, IEEE, 26(2), 77-83.
- Muir, B. M. 1994. Trust in automation: Part I. Theoretical issues in the study of trust and human intervention in automated sys- tems. Ergonomics, 37(11), 1905-1922.
- Muir, B. M., and Moray, N. 1996. Trust in automation. Part II. Experimental studies of trust and human intervention in a process control simulation. Ergonomics, 39(3), 429-460.
- Nielson, J. 1994. 10 Usability heuristics for user interface design. https://www.nngroup.com/articles/ten-usability-heuristics/ Naiseh, M., Jiang, N., Ma, J., & Ali, R. 2020. Personalising Ex- plainable Recommendations: Literature and Conceptualisation. In World Conference on Information Systems and Technologies (518-533). Springer, Cham.
- Ribera, M., and Lapedriza, A. 2019. Can we do better explana- tions? A proposal of user-centered explainable AI. In IUI Work- shops.
- Ribeiro, M. T., Singh, S., and Guestrin, C. 2016. Why Should I Trust You?: Explaining the Predictions of Any Classifier. In Pro- ceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 1135-1144). ACM. Rozenblit, L., and Keil, F. 2002. The misunderstood limits of folk science: An illusion of explanatory depth. Cognitive science, 26(5), 521-562.
- Rouse, W. B., and Morris, N. M. 1986. On looking into the black box: Prospects and limits in the search for mental models. Psy- chological Bulletin, 100(3), 349.
- Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. 2017. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE international conference on computer vision (pp. 618- 626).
- Settles, B. 2009. Active learning literature survey. University of Wisconsin-Madison Department of Computer Sciences.
- Sørmo, F., Cassens, J., and Aamodt, A. 2005. Explanation in case-based reasoning-perspectives and goals. Artificial Intelli- gence Review, 24(2), 109-143.
- Swartout, W. R. 1977. A digitalis therapy advisor with explana- tions. Proceedings of the 5th International Joint Conference on Artificial Intelligence-Volume 2, 819-825. http://dl.acm.org/cita- tion.cfm?id=1623009
- Swartout, W. R., and Moore, J. D. 1993. Explanation in second generation expert systems. Second Generation Expert Systems, 543-585.
- Tate, D., Grier, R., A. Martin, C., L. Moses, F., and Sparrow, D. 2016. A Framework for evidence based licensure of adaptive au- tonomous systems. IDA Paper P-5325. 10.13140/ RG.2.2.11845.86247
- Veinott, E., Klein, G., and Wiggins, S. 2010. Evaluating the ef- fect of the Premortem Method on Plan confidence. Paper pre- sented at the 7th International ISCRAM Conference 2010.
- Wang, D., Yang, Q., Abdul, A., & Lim, B. Y. 2019. Designing theory-driven user-centric explainable AI. In Proceedings of the 2019 CHI conference on human factors in computing systems (pp. 1-15).110111
- Wick, M. R., and Thompson, W. B. 1992. Reconstructive expert system explanation. Artificial Intelligence, 54(1-2), 33-70.
- Woolf, B. 2007. Building Intelligent Interactive Tutors: Student- centered strategies for revolutionizing e-learning. Morgan Kauf- mann Publishers Inc.
- Yang, S. C. H., and Shafto, P. 2017. Explainable artificial intelli- gence via Bayesian Teaching. In NIPS 2017 workshop on Teach- ing Machines, Robots, and Humans (pp. 127-137).
- Zeiler, M. D., and Fergus, R. 2014. Visualizing and understand- ing convolutional networks. In European conference on computer vision (pp. 818-833). Springer, Cham.