Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

From Word to Sense Embeddings: A Survey on Vector Representations of Meaning (1805.04032v3)

Published 10 May 2018 in cs.CL and cs.AI

Abstract: Over the past years, distributed semantic representations have proved to be effective and flexible keepers of prior knowledge to be integrated into downstream applications. This survey focuses on the representation of meaning. We start from the theoretical background behind word vector space models and highlight one of their major limitations: the meaning conflation deficiency, which arises from representing a word with all its possible meanings as a single vector. Then, we explain how this deficiency can be addressed through a transition from the word level to the more fine-grained level of word senses (in its broader acceptation) as a method for modelling unambiguous lexical meaning. We present a comprehensive overview of the wide range of techniques in the two main branches of sense representation, i.e., unsupervised and knowledge-based. Finally, this survey covers the main evaluation procedures and applications for this type of representation, and provides an analysis of four of its important aspects: interpretability, sense granularity, adaptability to different domains and compositionality.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
Citations (323)

Summary

A Survey on Vector Representations of Meaning: From Words to Senses

The paper "From Word to Sense Embeddings: A Survey on Vector Representations of Meaning" by Jose Camacho-Collados and Mohammad Taher Pilehvar provides an exhaustive survey on the landscape of semantic vector representations, extending from traditional word embeddings to the more granulated sense embeddings. The paper serves as a guide for researchers to understand both the historical context and recent advancements in modeling lexical semantics.

The work begins by addressing the well-recognized limitation of word embeddings, known as "meaning conflation deficiency". Traditional word embeddings are insufficient when handling polysemous words, as they represent all possible meanings of a word as a single vector. To navigate this deficiency, the paper examines the transition from words to word senses, mapping each meaning of a word to a distinct vector representation.

Two primary paradigms of sense representation are meticulously explored: unsupervised and knowledge-based approaches. Unsupervised methods derive sense distinctions from large corpora without the need for annotated data, making them adaptable but often lacking clarity in semantic distinctions. The survey highlights advancements such as joint embedding models that integrate sense induction and representation learning, and dynamic polysemy models that adapt to varying numbers of sense distinctions based on context. Contextualized word embeddings, which offer real-time adaptation to word meaning based on context, are also discussed as a subclass of unsupervised models, providing enhanced performance in various NLP tasks.

Conversely, knowledge-based methods capitalize on structured lexical resources like WordNet and BabelNet, offering higher interpretability and semantic accuracy. The paper provides a thorough review of these resources and details how sense and concept embeddings can be initialized and refined through textual definitions and semantic networks. The importance of these resources in bolstering word embeddings through processes like retrofitting is underscored, with applications ranging from text disambiguation to representation learning using knowledge base completion.

The survey also outlines the main evaluation methodologies, covering intrinsic evaluations such as word and sense similarity benchmarks, and extrinsic evaluations involving downstream tasks like sentiment analysis and word sense disambiguation. The discussion moves toward the practicality of these models, reflecting on the varying effectiveness of unsupervised versus knowledge-based techniques in real-world scenarios.

The implications of this research are profound, pushing the frontier towards a better semantic understanding in natural language processing. While significant strides have been made, there remain challenges such as the need for dynamic incorporation of sense nuances in evolving contexts and extending these approaches to multilingual settings. Future developments may see a bridge between discrete semantic inventories and continuous context-sensitive models, avoiding an explicit pre-disambiguation step—a direction that contextualized embeddings exemplify.

In conclusion, the paper offers a comprehensive roadmap for advancing semantic representation learning. It invites further exploration into dynamically integrating word meanings into AI applications, while grounding such advances in structured knowledge ultimately embodies the overarching aim of improving human-like language understanding in machines.