Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
156 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Interpretable Neural Embeddings with Sparse Self-Representation (2306.14135v1)

Published 25 Jun 2023 in cs.CL and cs.LG

Abstract: Interpretability benefits the theoretical understanding of representations. Existing word embeddings are generally dense representations. Hence, the meaning of latent dimensions is difficult to interpret. This makes word embeddings like a black-box and prevents them from being human-readable and further manipulation. Many methods employ sparse representation to learn interpretable word embeddings for better interpretability. However, they also suffer from the unstable issue of grouped selection in $\ell1$ and online dictionary learning. Therefore, they tend to yield different results each time. To alleviate this challenge, we propose a novel method to associate data self-representation with a shallow neural network to learn expressive, interpretable word embeddings. In experiments, we report that the resulting word embeddings achieve comparable and even slightly better interpretability than baseline embeddings. Besides, we also evaluate that our approach performs competitively well on all downstream tasks and outperforms benchmark embeddings on a majority of them.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. Linear algebraic structure of word senses, with applications to polysemy. Transactions of the Association of Computational Linguistics, 6:483–495.
  2. Tailoring continuous word representations for dependency parsing. In ACL (2), 809–815.
  3. Ehsan Elhamifar and René Vidal. 2009. Sparse subspace clustering. In Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on, pages 2790–2797. IEEE.
  4. Sparse overcomplete word vector representations. In Proc. of ACL.
  5. A closed form solution to robust subspace estimation and clustering. In Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on, pages 1801–1807. IEEE.
  6. Revisiting embedding features for simple semi-supervised learning. In EMNLP, 110–120.
  7. Fish transporters and miracle homes: How compositional distributional semantics can help np parsing. In Proc. of EMNLP.
  8. Fish transporters and miracle homes: How compositional distributional semantics can help np parsing. In Proceedings of EMNLP.
  9. X. Li and D Roth. 2006. Learning question classifiers: the role of semantic information. Natural Language Engineering, 12(03):229–249.
  10. Robust recovery of subspace structures by low-rank representation. IEEE transactions on pattern analysis and machine intelligence, 35(1):171–184.
  11. Robust and efficient subspace segmentation via least squares regression. In European conference on computer vision, pages 347–360. Springer.
  12. Online learning of interpretable word embeddings. Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, page 1687–1692.
  13. Distributed representations of words and phrases and their compositionality. In Advances in neural information processing systems, 3111–3119.
  14. Learning effective and interpretable semantic models using non-negative sparse embedding. In Proc. of COLING.
  15. Rotated word vector representations and their interpretability. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 401–411.
  16. Glove: Global vectors for word representation. In EMNLP, 14, 1532–43.
  17. Motion segmentation via robust subspace separation in the presence of outlying, incomplete, or corrupted trajectories.
  18. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the conference on empirical methods in natural language processing (EMNLP), 1631, 1642.
  19. Spine: Sparse interpretable neural embeddings. Proceedings of the Thirty Second AAAI Conference on Artificial Intelligence (AAAI).
  20. Sparse word embeddings using l1 regularized online learning. Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16).
  21. Robert Tibshirani. 1996. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society. Series B (Methodological), pages 267–288.
  22. Hui Zou and Trevor Hastie. 2005. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 67(2):301–320.

Summary

We haven't generated a summary for this paper yet.