Opening the black box: a primer for anti-discrimination
2021
Abstract
The pervasive adoption of Artificial Intelligence (AI) models in the modern information society, requires counterbalancing the growing decision power demanded to AI models with risk assessment methodologies. In this paper, we consider the risk of discriminatory decisions and review approaches for discovering discrimination and for designing fair AI models. We highlight the tight relations between discrimination discovery and explainable AI, with the latter being a more general approach for understanding the behavior of black boxes. SUMMARY: 1. AI risks. – 2. Discrimination discovery and fairness in AI. – 3. Explainable AI. – 4. Closing the gap. – 5. Conclusion.
Key takeaways
AI
AI
- AI models risk perpetuating discrimination due to historical biases in training data and decision-making processes.
- Discrimination discovery quantitatively assesses biases against protected groups using risk ratios, risk differences, and other measures.
- Explainable AI (XAI) enhances understanding of AI decisions, facilitating compliance with legal and ethical standards.
- Fairness in AI can be achieved through pre-processing, in-processing, post-processing, and run-time strategies.
- The interplay between discrimination discovery and XAI underscores the need for interpretable models to uncover hidden biases.
References (20)
- S. SAMOILI et al. Defining Artificial Intelligence. Towards an operational definition and taxonomy of Artificial Intelligence. EUR 30117 EN, Publications Office of the European Union, JRC118163, 2020.
- E. NTOUTSI et al. Bias in data-driven Artificial Intelligence systems -An introductory survey. In: Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 10.3, 2020.
- D. DANKS, A. J. LONDON. Regulating Autonomous Systems: Beyond Standards. In: IEEE Intell. Syst. 32.1, 2017, pp. 88-91.
- J. A. KROLL et al. Accountable Algorithms. In: U. of Penn. Law Review 165, 2017, pp. 633-705.
- Griggs v. Duke Power Co., 401 U.S. 424 (1971).
- Castaneda v. Partida, 430 U.S. 482 (1977).
- D. PEDRESCHI, S. RUGGIERI, F. TURINI. A study of top-k measures for discrimination discovery. In: SAC. ACM, 2012, pp. 126-131.
- L. ZHANG, Y. WU, X. WU. Causal Modeling-Based Discrimination Discovery and Removal: Criteria, Bounds, and Algorithms. In: IEEE Trans. Knowl. Data Eng. 31.11, 2019, pp. 2035-2050.
- A. CHOULDECHOVA, A. ROTH. A snapshot of the frontiers of fairness in machine learning. In: Commun. ACM 63.5, 2020, pp. 82-89.
- M. SRIVASTAVA, H. HEIDARI, A. KRAUSE. Mathematical Notions vs. Human Perception of Fairness: A Descriptive Approach to Fairness for Machine Learning. In: KDD. ACM, 2019, pp. 2459-2468.
- R. GUIDOTTI et al. A Survey of Methods for Explaining Black Box Models. In: ACM Comput. Surv. 51.5, 2019, 93:1-93:42.
- A. B. ARRIETA et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. in: Inf. Fusion 58, 2020, pp. 82-115.
- G. MALGIERI, G. COMANDE´. Why a Right to Legibility of Automated Decision-Making Exists in the GDPR. in: International Data Privacy Law 7.4, 2017, pp. 243-265.
- C. MOLNAR. Interpretable Machine Learning. Lulu.com, 2019.
- M. T. RIBEIRO, S. SINGH, C. GUESTRIN. "Why Should I Trust You?": Explaining the Predictions of Any Classifier. In: KDD. ACM, 2016, pp. 1135-1144.
- R. R. HOFFMAN, G. KLEIN. Explaining Explanation, Part 1: Theoretical Foundations. In: IEEE Intelligent Systems 32.3, 2017, pp. 68-73.
- T. MILLER. Explanation in Artificial Intelligence: Insights from the social sciences. In: Artificial Intelligence 267, 2019, pp. 1 -38.
- B. D. MITTELSTADT, C. RUSSELL, S. WACHTER. Explaining Explanations in AI. in: FAT. ACM, 2019, pp. 279-288.
- S. VENKATASUBRAMANIAN, M. ALFANO. The philosophical basis of algorithmic recourse. In: FAT*. ACM, 2020, pp. 284-293.
- R. GUIDOTTI et al. Factual and Counterfactual Explanations for Black Box Decision Making. In: IEEE Intell. Syst. 34.6, 2019, pp. 14-23.