Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Novel Neural Sequence Model with Multiple Attentions for Word Sense Disambiguation (1809.01074v1)

Published 4 Sep 2018 in cs.CL and cs.LG

Abstract: Word sense disambiguation (WSD) is a well researched problem in computational linguistics. Different research works have approached this problem in different ways. Some state of the art results that have been achieved for this problem are by supervised models in terms of accuracy, but they often fall behind flexible knowledge-based solutions which use engineered features as well as human annotators to disambiguate every target word. This work focuses on bridging this gap using neural sequence models incorporating the well-known attention mechanism. The main gist of our work is to combine multiple attentions on different linguistic features through weights and to provide a unified framework for doing this. This weighted attention allows the model to easily disambiguate the sense of an ambiguous word by attending over a suitable portion of a sentence. Our extensive experiments show that multiple attention enables a more versatile encoder-decoder model leading to state of the art results.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Mahtab Ahmed (5 papers)
  2. Muhammad Rifayat Samee (4 papers)
  3. Robert E. Mercer (14 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.