Academia.eduAcademia.edu

Model Transparency

description17 papers
group0 followers
lightbulbAbout this topic
Model transparency refers to the degree to which the internal workings and decision-making processes of a model, particularly in machine learning and artificial intelligence, are understandable and interpretable by humans. It emphasizes clarity in how models operate, enabling stakeholders to assess their reliability, fairness, and accountability.
lightbulbAbout this topic
Model transparency refers to the degree to which the internal workings and decision-making processes of a model, particularly in machine learning and artificial intelligence, are understandable and interpretable by humans. It emphasizes clarity in how models operate, enabling stakeholders to assess their reliability, fairness, and accountability.

Key research themes

1. How can standardized frameworks define and assess transparency levels for diverse stakeholders in autonomous and AI systems?

This research area focuses on developing measurable, testable standards to specify and assess transparency in autonomous systems, addressing the varying needs of different stakeholders such as users, regulators, and investigators. Establishing such frameworks matters to ensure accountability, trust, and safety by making AI systems understandable and their decisions explicable across multiple application contexts.

Key finding: Introduces IEEE P7001 draft standard as a structured approach defining testable transparency levels tailored to five stakeholder groups (users, public/bystanders, safety agencies, investigators, and lawyers). The standard... Read more
Key finding: Develops the concept of Transparency by Design (TbD) integrating contextual, technical, informational, and stakeholder-sensitive principles into AI system development. Proposes nine principles inspired by privacy-by-design... Read more
Key finding: Proposes Method Cards as prescriptive documentation artifacts that go beyond descriptive transparency by providing actionable guidance to ML engineers on model reproduction, design rationales, and mitigation strategies for... Read more

2. What are the epistemic and practical challenges underlying transparency in complex AI-driven simulations and computational systems?

This research theme investigates the nature of opacity and transparency in complex computational systems such as AI, computer simulations, and big data applications. It explores the conceptual limits of knowledge and understanding about system internals, addresses the multiple layers of opacity, and evaluates how partial or instrumental transparency can be attained to support scientific explanations, artifact detection, and trustworthy deployment.

Key finding: Reconceptualizes opacity beyond Humphreys’ computational steps inaccessible by hand, defining opacity as the disposition to resist epistemic access including forms of knowledge and understanding. It distinguishes different... Read more
Key finding: Analyzes transparency as consisting of three forms—functional transparency (algorithmic functioning), structural transparency (implementation in code), and run transparency (actual execution on hardware and data)—to address... Read more

3. How do methodological and user-centered approaches advance transparency in interpretability, data documentation, and explanation of AI models?

This area studies practical frameworks and methodologies to improve transparency through structured documentation, interpretability techniques, and user-centric explanations. It focuses on additive versus non-additive model explanations, transparent documentation of datasets and processes, and enhanced user communication to bridge the gap between technical AI design and interpretability by diverse stakeholders.

Key finding: Compares multiple additive explanation methods (partial dependence, Shapley explanations, distilled additive explanations, gradient-based explanations) for black-box models, revealing that distilled additive explanations... Read more
Key finding: Explores transparency as a situated and evolving process in qualitative research methodology, emphasizing 'methodological data' as reflexive artifacts that complicate simplistic accounts of transparency. Argues that... Read more
Key finding: Proposes Method Cards as a novel documentation tool for machine learning that combines descriptive information with prescriptive guidance. These cards facilitate model reproduction, understand design choices, and provide... Read more

All papers in Model Transparency

The interpretability of deep neural networks (DNNs) is a critical focus in artificial intelligence (AI) and machine learning (ML), particularly as these models are increasingly deployed in high-stakes applications such as healthcare,... more
In the natural and social sciences, it is common to use toy models-extremely simple and highly idealized representations-to understand complex phenomena. Some of the simple surrogate models used to understand opaque machine learning (ML)... more
Motivated by a recent trend that advocates a reassessment of the aim of medical science and clinical practice, this paper investigates the epistemic aims of biomedical research. Drawing on contemporary discussions in epistemology and the... more
Consistency, scalability, and local stability properties ensure that a model or method produces reliable and predictable outcomes. The Shapash helps users understand how the model makes its decisions. With machine learning (ML) system,... more
This study investigates the Shapley additive explanation (SHAP) of the extreme boosting (XGBoost) model for breast cancer diagnosis. The study employed Wisconsin’s breast cancer dataset, characterized by 30 features extracted from an... more
This study investigates the Shapley additive explanation (SHAP) of the extreme boosting (XGBoost) model for breast cancer diagnosis. The study employed Wisconsin’s breast cancer dataset, characterized by 30 features extracted from an... more
In a recent paper, Erasmus et al. (2021) defend the idea that the ambiguity of the term "explanation" in explainable AI (XAI) can be solved by adopting any of four different extant accounts of explanation in the philosophy of science: the... more
This is an integrative review that address the question, "What makes for a good explanation?" with reference to AI systems. Pertinent literatures are vast. Thus, this review is necessarily selective. That said, most of the key... more
This study investigates the Shapley additive explanation (SHAP) of the extreme boosting (XGBoost) model for breast cancer diagnosis. The study employed Wisconsin’s breast cancer dataset, characterized by 30 features extracted from an... more
This study investigates the Shapley additive explanation (SHAP) of the extreme boosting (XGBoost) model for breast cancer diagnosis. The study employed Wisconsin's breast cancer dataset, characterized by 30 features extracted from an... more
Previous research in Explainable Artificial Intelligence (XAI) suggests that a main aim of explainability approaches is to satisfy specific interests, goals, expectations, needs, and demands regarding artificial systems (we call these... more
Power system phasor time-domain simulation is often carried out through domain specific tools such as Eurostag, PSS/E, and others. While these tools are efficient, their individual sub-component models and solvers cannot be accessed by... more
In order for a machine learning model to be useful, it must be used. Opaque models that predict or classify without explaining are often ignored. Thus measuring the satisfaction of those who receive an explanation is one natural way to... more
This is an integrative review that address the question, "What makes for a good explanation?" with reference to AI systems. Pertinent literatures are vast. Thus, this review is necessarily selective. That said, most of the key... more
In the paper, I review some of the emerging philosophical literature on the problem of using artificial neural networks (ANNs) and deep learning in science. Specifically, I focus on the problem of opacity in such systems and argue that... more
The field of Artificial Intelligence has seen dramatic progress over the last 15 years. Using machine learning methods, software systems that automatically learn and improve relationships using digitized experience, researchers and... more
Supervised machine-learning models boast remarkable predictive capabilities. But can you trust your model? Will it work in deployment? What else can it tell you about the world?
Recent work on interpretability in machine learning and AI has focused on the building of simplified models that approximate the true criteria used to make decisions. These models are a useful pedagogical device for teaching trained... more
Explainability and comprehensibility of AI are important requirements for intelligent systems deployed in real-world domains. Users want and frequently need to understand how decisions impacting them are made. Similarly it is important to... more
Previous research in Explainable Artificial Intelligence (XAI) suggests that a main aim of explainability approaches is to satisfy specific interests, goals, expectations, needs, and demands regarding artificial systems (we call these... more
The rapid growth of research in explainable artificial intelligence (XAI) follows on two substantial developments. First, the enormous application success of modern machine learning methods, especially deep and reinforcement learning,... more
Many high-performance models suffer from a lack of interpretability. There has been an increasing influx of work on explainable artificial intelligence (XAI) in order to disentangle what is meant and expected by XAI. Nevertheless, there... more
The field of explainable AI (XAI) has quickly become a thriving and prolific community. However, a silent, recurrent and acknowledged issue in this area is the lack of consensus regarding its terminology. In particular, each new... more
The field of explainable AI (XAI) has quickly become a thriving and prolific community. However, a silent, recurrent and acknowledged issue in this area is the lack of consensus regarding its terminology. In particular, each new... more
Explainability is assumed to be a key factor for the adoption of Artificial Intelligence systems in a wide range of contexts (Hoffman, Mueller, & Klein, 2017; Hoffman, Mueller, Klein, & Litman, 2018; Doran, Schulz, & Besold, 2017; Lipton,... more
Copyright held by the owner/author(s). CHI’21, May 8-13, 2021, Online Virtual Conference ACM 978-1-4503-6819-3/20/04. https://doi.org/10.1145/3334480.XXXXXXX Abstract Given that there are a variety of stakeholders involved in, and... more
The rapid growth of research in explainable artificial intelligence (XAI) follows on two substantial developments. First, the enormous application success of modern machine learning methods, especially deep and reinforcement learning,... more
The widespread adoption of black-box models in Artificial Intelligence has enhanced the need for explanation methods to reveal how these obscure models reach specific decisions. Retrieving explanations is fundamental to unveil possible... more
by Giulia Vilone and 
1 more
Explainable Artificial Intelligence (XAI) has experienced a significant growth over the last few years. This is due to the widespread application of machine learning, particularly deep learning, that has led to the development of highly... more
The technologies supporting Artificial Intelligence (AI) have advanced rapidly over the past few years and AI is becoming a commonplace in every aspect of life like the future of self-driving cars or earlier health diagnosis. For this to... more
We argue that artificial networks are explainable and offer a novel theory of interpretability. Two sets of conceptual questions are prominent in theoretical engagements with artificial neural networks, especially in the context of... more
In this paper I argue that the search for explainable models and interpretable decisions in AI must be reformulated in terms of the broader project of offering a pragmatic and naturalistic account of understanding in AI. Intuitively, the... more
Recent work on interpretability in machine learning and AI has focused on the building of simplified models that approximate the true criteria used to make decisions. These models are a useful pedagogical device for teaching trained... more
Download research papers for free!