Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

An Investigation on Word Embedding Offset Clustering as Relationship Classification (2305.04265v1)

Published 7 May 2023 in cs.CL

Abstract: Vector representations obtained from word embedding are the source of many groundbreaking advances in natural language processing. They yield word representations that are capable of capturing semantics and analogies of words within a text corpus. This study is an investigation in an attempt to elicit a vector representation of relationships between pairs of word vectors. We use six pooling strategies to represent vector relationships. Different types of clustering models are applied to analyze which one correctly groups relationship types. Subtraction pooling coupled with a centroid based clustering mechanism shows better performances in our experimental setup. This work aims to provide directions for a word embedding based unsupervised method to identify the nature of a relationship represented by a pair of words.

Summary

We haven't generated a summary for this paper yet.