2000 character limit reached
Robust Counterfactual Explanations in Machine Learning: A Survey (2402.01928v1)
Published 2 Feb 2024 in cs.LG and cs.AI
Abstract: Counterfactual explanations (CEs) are advocated as being ideally suited to providing algorithmic recourse for subjects affected by the predictions of machine learning models. While CEs can be beneficial to affected individuals, recent work has exposed severe issues related to the robustness of state-of-the-art methods for obtaining CEs. Since a lack of robustness may compromise the validity of CEs, techniques to mitigate this risk are in order. In this survey, we review works in the rapidly growing area of robust CEs and perform an in-depth analysis of the forms of robustness they consider. We also discuss existing solutions and their limitations, providing a solid foundation for future developments.
- Evaluating robustness of counterfactual explanations. In IEEE SSCI, 2021.
- Selective ensembles for consistent predictions. In ICLR, 2022.
- Model multiplicity: Opportunities, concerns, and solutions. In FAccT, 2022.
- Consistent counterfactuals for deep models. In ICLR, 2022.
- Leo Breiman. Statistical modeling: The two cultures. Statistical science, 16(3):199–231, 2001.
- Counterfactual plans under distributional ambiguity. In ICLR, 2022.
- Coverage-validity-aware algorithmic recourse. arXiv:2311.11349, 2023.
- Characterizing fairness over the set of good models under selective labels. In ICML, 2021.
- Using the k-associated optimal graph to provide counterfactual explanations. In FUZZ-IEEE, 2022.
- On the adversarial robustness of causal algorithmic recourse. In ICML, 2022.
- Robust counterfactual explanations for tree-based ensembles. In ICML, 2022.
- Robustness implies fairness in causal algorithmic recourse. In FAccT, 2023.
- The robustness of counterfactual explanations over time. IEEE Access, 10:82736–82750, 2022.
- Setting the right expectations: Algorithmic recourse over time. In EAAMO, 2023.
- Robust counterfactual explanations for random forests. arXiv:2205.14116, 2022.
- Rocoursenet: Robust training of a prediction aware recourse model. In CIKM, 2023.
- Equalizing recourse across groups. arXiv:1909.03166, 2019.
- Generating robust counterfactual explanations. In ECML PKDD, 2023.
- Exploring counterfactual explanations for classification and regression trees. In ECML PKDD Workshops, 2021.
- Robust counterfactual explanations for neural networks with probabilistic guarantees. In ICML, 2023.
- Leif Hancox-Li. Robustness in machine learning explanations: does it matter? In FAT*, 2020.
- Provably robust and plausible counterfactual explanations for neural networks via robust optimisation. In ACML, 2023.
- Formalising the robustness of counterfactual explanations for neural networks. In AAAI, 2023.
- Recourse under model multiplicity via argumentative ensembling. In AAMAS, 2024.
- Counterfactual explanation with missing values. arXiv:2304.14606, 2023.
- A survey of algorithmic recourse: Contrastive explanations and consequential recommendations. ACM CSUR, 55(5):1–29, 2022.
- On the impact of adversarially robust models on algorithmic recourse. In NeurIPS Workshops, 2022.
- Towards bridging the gaps between the right to explanation and the right to be forgotten. In ICML, 2023.
- Robust explanations for human-neural multi-agent systems with formal verification. In EUMAS, 2023.
- Promoting counterfactual robustness through diversity. In AAAI, 2024.
- Counterfactual explanations and model multiplicity: a relational verification view. In KR, 2023.
- Finding regions of counterfactual explanations via robust optimization. arXiv:2301.11113, 2023.
- Predictive multiplicity in classification. In ICML, 2020.
- A survey on the robustness of feature importance and counterfactual explanations. arXiv:2111.00358, 2021.
- Robust explanations for private support vector machines. arXiv:2102.03785, 2021.
- Scaling guarantees for nearest counterfactual explanations. In AIES, 2021.
- Robust bayesian recourse. In UAI, 2022.
- Distributionally robust recourse action. In ICLR, 2023.
- On counterfactual explanations under predictive multiplicity. In UAI, 2020.
- CARLA: A python library to benchmark algorithmic recourse and counterfactual explanation algorithms. In NeurIPS Datasets and Benchmarks, 2021.
- Exploring counterfactual explanations through the lens of adversarial examples: A theoretical and empirical analysis. In AISTATS, 2022.
- Probabilistically robust recourse: Navigating the trade-offs between costs and robustness in algorithmic recourse. In ICLR, 2023.
- On the trade-off between actionable explanations and the right to be forgotten. In ICLR, 2023.
- Bayesian hierarchical models for counterfactual estimation. In AISTATS, 2023.
- Algorithmic recourse in the wild: Understanding the impact of data and model shifts. arXiv:2012.11788, 2020.
- "why should I trust you?": Explaining the predictions of any classifier. In KDD, 2016.
- Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell., 1(5):206–215, 2019.
- CERTIFAI: A common framework to provide explanations and analyse the fairness and robustness of black-box models. In AIES, 2020.
- FASTER-CE: fast, sparse, transparent, and robust counterfactual explanations. arXiv:2210.06578, 2022.
- Counterfactual explanations can be manipulated. In NeurIPS, 2021.
- Interpretable predictions of tree-based ensembles via actionable feature tweaking. In KDD, 2017.
- Towards robust and reliable algorithmic recourse. In NeurIPS, 2021.
- On the robustness of sparse counterfactual explanations to adverse perturbations. Artif. Intell., 316:103840, 2023.
- On the fairness of causal algorithmic recourse. In AAAI, 2022.
- Counterfactual explanations without opening the black box: Automated decisions and the GDPR. Harv. JL & Tech., 31:841, 2017.
- T-COL: generating counterfactual explanations for general user preferences on variable machine learning systems. arXiv:2309.16146, 2023.
- Flexible and robust counterfactual explanations with minimal satisfiable perturbations. In CIKM, 2023.
- Density-based reliable and robust explainer for counterfactual explanation. Expert Syst. Appl., 226:120214, 2023.