2000 character limit reached
Towards Benchmarking the Utility of Explanations for Model Debugging (2105.04505v1)
Published 10 May 2021 in cs.AI, cs.HC, and cs.LG
Abstract: Post-hoc explanation methods are an important class of approaches that help understand the rationale underlying a trained model's decision. But how useful are they for an end-user towards accomplishing a given task? In this vision paper, we argue the need for a benchmark to facilitate evaluations of the utility of post-hoc explanation methods. As a first step to this end, we enumerate desirable properties that such a benchmark should possess for the task of debugging text classifiers. Additionally, we highlight that such a benchmark facilitates not only assessing the effectiveness of explanations but also their efficiency.
- Maximilian Idahl (5 papers)
- Lijun Lyu (6 papers)
- Ujwal Gadiraju (28 papers)
- Avishek Anand (81 papers)