Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ConceptDistil: Model-Agnostic Distillation of Concept Explanations (2205.03601v1)

Published 7 May 2022 in cs.LG and cs.AI

Abstract: Concept-based explanations aims to fill the model interpretability gap for non-technical humans-in-the-loop. Previous work has focused on providing concepts for specific models (eg, neural networks) or data types (eg, images), and by either trying to extract concepts from an already trained network or training self-explainable models through multi-task learning. In this work, we propose ConceptDistil, a method to bring concept explanations to any black-box classifier using knowledge distillation. ConceptDistil is decomposed into two components:(1) a concept model that predicts which domain concepts are present in a given instance, and (2) a distillation model that tries to mimic the predictions of a black-box model using the concept model predictions. We validate ConceptDistil in a real world use-case, showing that it is able to optimize both tasks, bringing concept-explainability to any black-box model.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. João Bento Sousa (1 paper)
  2. Ricardo Moreira (5 papers)
  3. Vladimir Balayan (5 papers)
  4. Pedro Saleiro (39 papers)
  5. Pedro Bizarro (41 papers)
Citations (3)