A SURVEY ON EXPLAINABLE AI: TECHNIQUES AND CHALLENGES
2020, Novateur Publications, International Journal of Innovations in Engineering Research and Technology.
Abstract
Explainable Artificial Intelligence (XAI) is a rapidly evolving field aimed at making AI systems more interpretable and transparent to human users. As AI technologies become increasingly integrated into critical sectors such as healthcare, finance, and autonomous systems, the need for explanations behind AI decisions has grown significantly. This survey provides a comprehensive review of XAI techniques, categorizing them into post-hoc and intrinsic methods, and examines their application in various domains. Additionally, the paper explores the major challenges in achieving explainability, including balancing accuracy with interpretability, scalability, and the trade-off between transparency and complexity. The survey concludes with a discussion on the future directions of XAI, emphasizing the importance of interdisciplinary approaches to developing robust and interpretable AI systems.
FAQs
AI
What explains the trade-off between model accuracy and interpretability in XAI techniques?
The analysis highlights that deep neural networks are highly accurate but less interpretable, while simpler models like decision trees offer high interpretability at the cost of accuracy, resulting in an average accuracy reduction of 15-20%.
How did LIME and SHAP compare in terms of interpretability and fidelity?
Both LIME and SHAP provide high interpretability, scoring around 2-5% accuracy loss, with SHAP exhibiting higher fidelity values between 0.8 to 0.9 due to its global explanation capabilities.
When was SHAP introduced, and what is its significance in XAI?
Introduced by Lundberg and Lee in 2017, SHAP uses cooperative game theory to provide a consistent framework for feature attribution, making it one of the most widely adopted tools in XAI.
What challenges do intrinsic XAI methods face in complex tasks?
Intrinsic methods like decision trees maintain high interpretability but require significant simplifications, leading to a 15-20% accuracy reduction in complex tasks such as image recognition compared to more complex models.
Why is audience-specific explanation crucial in XAI frameworks?
The need for audience-specific explanations arises from the varying levels of technical understanding; explanations that are interpretable for data scientists may confuse non-technical users, hindering trust and usability.
References (10)
- Doshi-Velez, F., & Kim, B. (2017). Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608.
- Gilpin, L. H., Bau, D., Yuan, B. Z., Bajwa, A., Specter, M., & Kagal, L. (2018). Explaining explanations: An overview of interpretability of machine learning. 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA), 80-89.
- Lipton, Z. C. (2016). The mythos of model interpretability. arXiv preprint arXiv:1606.03490.
- Ribeiro, M. T., Singh, S., & Guestrin, C. (2016). "Why should I trust you?" Explaining the predictions of any classifier. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 1135-1144.
- Samek, W., Wiegand, T., & Müller, K. R. (2017). Explainable artificial intelligence: Understanding, visualizing and interpreting deep learning models. arXiv preprint arXiv:1708.08296.
- Shrikumar, A., Greenside, P., & Kundaje, A. (2017). Learning important features through propagating activation differences. Proceedings of the 34th International Conference on Machine Learning, 3145- 3153
- Sundararajan, M., Taly, A., & Yan, Q. (2017). Axiomatic attribution for deep networks. Proceedings of the 34th International Conference on Machine Learning, 3319-3328.
- Tjoa, E., & Guan, C. (2019). A survey on explainable artificial intelligence (XAI): Towards medical XAI. IEEE Transactions on Neural Networks and Learning Systems, 32(11), 4793-4813.
- Montavon, G., Samek, W., & Müller, K. R. (2018). Methods for interpreting and understanding deep neural networks. Digital Signal Processing, 73, 1-15.
- Letham, B., Rudin, C., McCormick, T. H., & Madigan, D. (2015). Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model. The Annals of Applied Statistics, 9(3), 1350- 1371.