Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Many Faces of Feature Importance: Comparing Built-in and Post-hoc Feature Importance in Text Classification (1910.08534v1)

Published 18 Oct 2019 in cs.CL, cs.CY, cs.HC, and cs.LG

Abstract: Feature importance is commonly used to explain machine predictions. While feature importance can be derived from a machine learning model with a variety of methods, the consistency of feature importance via different methods remains understudied. In this work, we systematically compare feature importance from built-in mechanisms in a model such as attention values and post-hoc methods that approximate model behavior such as LIME. Using text classification as a testbed, we find that 1) no matter which method we use, important features from traditional models such as SVM and XGBoost are more similar with each other, than with deep learning models; 2) post-hoc methods tend to generate more similar important features for two models than built-in methods. We further demonstrate how such similarity varies across instances. Notably, important features do not always resemble each other better when two models agree on the predicted label than when they disagree.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Vivian Lai (28 papers)
  2. Jon Z. Cai (5 papers)
  3. Chenhao Tan (89 papers)
Citations (18)