Explainable Machine Learning in Deployment: An Overview
The paper "Explainable Machine Learning in Deployment" by Bhatt et al. provides an in-depth exploration of the practical deployment of explainable ML techniques in various organizational contexts. It aims to bridge the gap between explainability methodologies developed in academic settings and their real-world applications. The research focuses on understanding how explainability is viewed and utilized by organizations, particularly emphasizing the difference between internal and external stakeholders.
Key Findings and Techniques
The paper's primary focus is on local explainability methods, highlighting their practical deployment across different industries. It identifies a pronounced gap between the theoretical promise of explainability and its actual application, which tends to cater more to internal stakeholders, like ML engineers, rather than external end users. The paper identifies several key methodologies, each with specific insights:
- Feature Importance: This is the most commonly deployed explainability technique. It often uses Shapley values to provide insights into the significance of individual features in the prediction process. The authors note that while this technique is widely used to sanity-check model outputs, it rarely serves to explain predictions to end users.
- Counterfactual Explanations: These explanations help understand model outputs by identifying minimal changes to input data that would change the model's prediction. The paper identifies these as potential tools for providing recourse, although practical deployment faces challenges due to computational constraints and the need for plausibility.
- Adversarial Training: This method improves model robustness and explainability by focusing on features that are consistent across adversarial examples. The paper remarks on the surprising correlation between robustness and interpretability, offering insights into improving ML model reliability.
- Influential Samples: Techniques like influence functions attempt to identify which training data points most affect a given prediction. Despite theoretical interest, practical deployment is limited due to computational and interpretational challenges, particularly in handling outliers.
Methodology
The paper synthesizes insights from approximately fifty interviews with stakeholders from around thirty organizations, including both non-profit and for-profit entities. It categorizes stakeholders into several groups – executives, ML engineers, end users, and other stakeholders – to dissect their specific needs and engagements with explainability.
Implications and Recommendations
The paper suggests that many organizations still rely heavily on domain experts to filter and validate explanations, indicating a mismatch between technical capabilities and practical needs. It stresses the importance of setting clear goals for explainability, recommending a structured framework to establish stakeholder-specific desiderata.
Furthermore, the paper raises concerns about explainability, such as privacy issues, the challenge of ensuring causal rather than correlative explanations, and the dual-use nature of improved model understanding, which can also empower malicious applications.
Future Directions
The research highlights several avenues for future work, emphasizing the need for:
- Causal Explanations: Developing methods that provide causal insights as opposed to merely correlative ones.
- Scalable Solutions: Addressing computational inefficiencies to enable real-time explainability.
- Framework Development: Creating frameworks that align explanation techniques with specific organizational goals and contexts.
- Regulatory and Ethical Considerations: Navigating the evolving legal landscape regarding explainability mandates.
This paper provides a comprehensive view that not only documents current practices but also stimulates further research and technological advancements in the field of explainable AI.