Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
133 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Intrinsic User-Centric Interpretability through Global Mixture of Experts (2402.02933v4)

Published 5 Feb 2024 in cs.LG, cs.CY, and cs.HC

Abstract: In human-centric settings like education or healthcare, model accuracy and model explainability are key factors for user adoption. Towards these two goals, intrinsically interpretable deep learning models have gained popularity, focusing on accurate predictions alongside faithful explanations. However, there exists a gap in the human-centeredness of these approaches, which often produce nuanced and complex explanations that are not easily actionable for downstream users. We present InterpretCC (interpretable conditional computation), a family of intrinsically interpretable neural networks at a unique point in the design space that optimizes for ease of human understanding and explanation faithfulness, while maintaining comparable performance to state-of-the-art models. InterpretCC achieves this through adaptive sparse activation of features before prediction, allowing the model to use a different, minimal set of features for each instance. We extend this idea into an interpretable, global mixture-of-experts (MoE) model that allows users to specify topics of interest, discretely separates the feature space for each data point into topical subnetworks, and adaptively and sparsely activates these topical subnetworks for prediction. We apply InterpretCC for text, time series and tabular data across several real-world datasets, demonstrating comparable performance with non-interpretable baselines and outperforming intrinsically interpretable baselines. Through a user study involving 56 teachers, InterpretCC explanations are found to have higher actionability and usefulness over other intrinsically interpretable approaches.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. GPT-4 technical report. arXiv preprint arXiv:2303.08774, 2023.
  2. Ripple: Concept-based interpretation for raw time series models in education. In Proceedings of the 13th AAAI Symposium on Educational Advances in Artificial Intelligence, 2022. Accepted as a full paper at AAAI 2023: 37th AAAI Conference on Artificial Intelligence (EAAI: AI for Education Special Track), 7-14 of February 2023, Washington DC, USA.
  3. Ripple: Concept-based interpretation for raw time series models in education. In Proceedings of the AAAI Conference on Artificial Intelligence, 2023.
  4. Conditional computation in neural networks for faster models, 2015.
  5. Estimating or propagating gradients through stochastic neurons for conditional computation, 2013.
  6. How to quantify student’s regularity? In Verbert, K., Sharples, M., and Klobučar, T. (eds.), Adaptive and Adaptable Learning, pp.  277–291, Cham, 2016. Springer International Publishing. ISBN 978-3-319-45153-4.
  7. This looks like that: deep learning for interpretable image recognition. Advances in neural information processing systems, 32, 2019.
  8. Utilizing student time series behaviour in learning management systems for early prediction of course performance. Journal of Learning Analytics, 7(2):1–17, 9 2020.
  9. Low-rank approximations for conditional feedforward computation in deep neural networks, 2013.
  10. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
  11. Explanations based on the missing: Towards contrastive explanations with pertinent negatives. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, pp.  590–601, Red Hook, NY, USA, 2018. Curran Associates Inc.
  12. A review of sparse expert models in deep learning. arXiv preprint arXiv:2209.01667, 2022a.
  13. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. The Journal of Machine Learning Research, 23(1):5232–5270, 2022b.
  14. Towards automatic concept-based explanations, 2019.
  15. Measuring student’s utilization of video resources and its effect on academic performance. In 2018 IEEE 18th International Conference on Advanced Learning Technologies (ICALT), pp.  196–198, 2018.
  16. HF Canonical Model Maintainers. distilbert-base-uncased-finetuned-sst-2-english (revision bfdd146), 2022.
  17. Interpretable mixture of experts. Transactions on Machine Learning Research, 2023.
  18. Categorical reparameterization with gumbel-softmax. In International Conference on Learning Representations, 2017.
  19. Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024.
  20. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). ICML 2018, 2017.
  21. The disagreement problem in explainable machine learning: A practitioner’s perspective. arXiv preprint arXiv:2202.01602, 2022.
  22. A data-driven student model to provide adaptive support during video watching across moocs. In Bittencourt, I. I., Cukurova, M., Muldner, K., Luckin, R., and Millán, E. (eds.), Artificial Intelligence in Education, pp.  282–295, Cham, 2020. Springer International Publishing. ISBN 978-3-030-52237-7.
  23. Grade prediction of weekly assignments in moocs: mining video-viewing behavior. Education and Information Technologies, 25(2):1333–1342, 2020.
  24. Base layers: Simplifying training of large, sparse models. In International Conference on Machine Learning, pp.  6265–6274. PMLR, 2021.
  25. Branch-train-merge: Embarrassingly parallel training of expert language models. arXiv preprint arXiv:2208.03306, 2022.
  26. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, pp.  4768–4777, Red Hook, NY, USA, 2017. Curran Associates Inc. ISBN 9781510860964.
  27. Can feature predictive power generalize? benchmarking early predictors of student success across flipped and online courses. Proceedings of the 14th International Conference on Educational Data Mining, 2021.
  28. Early prediction of success in mooc from video interaction features. In Bittencourt, I. I., Cukurova, M., Muldner, K., Luckin, R., and Millán, E. (eds.), Artificial Intelligence in Education, pp.  191–196, Cham, 2020. Springer International Publishing. ISBN 978-3-030-52240-7.
  29. Identifying and comparing multi-dimensional student profiles across flipped classrooms. In Artificial Intelligence in Education Proceedings, volume Part I of Lecture Notes in Computer Science. 13355, pp.  90–102. Springer, 2022.
  30. Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence, 267:1–38, 2019.
  31. Predictive learning analytics using deep learning model in moocs’ courses videos. Education and Information Technologies, 26:371–392, 2021.
  32. Multimodal contrastive learning with limoe: the language-image mixture of experts. Advances in Neural Information Processing Systems, 35:9564–9576, 2022.
  33. Pip-net: Patch-based intuitive prototypes for interpretable image classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp.  2744–2753, 2023.
  34. Sentence-bert: Sentence embeddings using siamese bert-networks. arXiv preprint arXiv:1908.10084, 2019.
  35. ”why should i trust you?”: Explaining the predictions of any classifier. 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp.  1135–1144, 2016.
  36. Satija, M. P. The theory and practice of the Dewey decimal classification system. Elsevier, 2013.
  37. Concept bottleneck model with additional unsupervised concepts. IEEE Access, 10:41758–41765, 2022.
  38. Dewey decimal classification. Libraries Unlimited, 1998.
  39. Grad-CAM: Visual explanations from deep networks via gradient-based localization. International Journal of Computer Vision, 128(2):336–359, 10 2019.
  40. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pp.  1631–1642, 2013.
  41. Meta transfer learning for early success prediction in moocs. In Proceedings of the Ninth ACM Conference on Learning@ Scale, pp.  121–132, 2022a.
  42. Meta transfer learning for early success prediction in moocs, 2022b.
  43. Evaluating the explainers: Black-box explainable machine learning for student success prediction in MOOCs. In Mitrovic, A. and Bosch, N. (eds.), Proceedings of the 15th International Conference on Educational Data Mining, pp.  98–109, Durham, United Kingdom, 7 2022c. International Educational Data Mining Society. ISBN 978-1-7336736-3-1.
  44. Trusting the explainers: teacher validation of explainable artificial intelligence for course design. In LAK23: 13th International Learning Analytics and Knowledge Conference, pp.  345–356, 2023a.
  45. The future of human-centric explainable artificial intelligence (xai) is not post-hoc explanations. arXiv preprint arXiv:2307.00364, 2023b.
  46. Dynamic convolutions: Exploiting spatial sparsity for faster inference, 2019.
  47. Pedagogical intervention practices: Improving learning engagement based on early prediction. IEEE Transactions on Learning Technologies, 12(2):278–289, 2019.
  48. Breast Cancer Wisconsin (Diagnostic). UCI Machine Learning Repository, 1995. DOI: https://doi.org/10.24432/C5DW2B.
  49. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems, 32, 2019.
  50. Sum-of-parts models: Faithful attributions for groups of features. arXiv preprint arXiv:2310.16316, 2023.
  51. Character-level convolutional networks for text classification. Advances in neural information processing systems, 28, 2015.
  52. Designing effective sparse expert models. arXiv preprint arXiv:2202.08906, 2, 2022.
Citations (3)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com