Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Low Anisotropy Sense Retrofitting (LASeR) : Towards Isotropic and Sense Enriched Representations (2104.10833v1)

Published 22 Apr 2021 in cs.CL

Abstract: Contextual word representation models have shown massive improvements on a multitude of NLP tasks, yet their word sense disambiguation capabilities remain poorly explained. To address this gap, we assess whether contextual word representations extracted from deep pretrained LLMs create distinguishable representations for different senses of a given word. We analyze the representation geometry and find that most layers of deep pretrained LLMs create highly anisotropic representations, pointing towards the existence of representation degeneration problem in contextual word representations. After accounting for anisotropy, our study further reveals that there is variability in sense learning capabilities across different LLMs. Finally, we propose LASeR, a 'Low Anisotropy Sense Retrofitting' approach that renders off-the-shelf representations isotropic and semantically more meaningful, resolving the representation degeneration problem as a post-processing step, and conducting sense-enrichment of contextualized representations extracted from deep neural LLMs.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Geetanjali Bihani (10 papers)
  2. Julia Taylor Rayz (17 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.