Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Gating Mechanisms for Combining Character and Word-level Word Representations: An Empirical Study (1904.05584v1)

Published 11 Apr 2019 in cs.CL and stat.ML

Abstract: In this paper we study how different ways of combining character and word-level representations affect the quality of both final word and sentence representations. We provide strong empirical evidence that modeling characters improves the learned representations at the word and sentence levels, and that doing so is particularly useful when representing less frequent words. We further show that a feature-wise sigmoid gating mechanism is a robust method for creating representations that encode semantic similarity, as it performed reasonably well in several word similarity datasets. Finally, our findings suggest that properly capturing semantic similarity at the word level does not consistently yield improved performance in downstream sentence-level tasks. Our code is available at https://github.com/jabalazs/gating

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (2)
  1. Jorge A. Balazs (7 papers)
  2. Yutaka Matsuo (128 papers)
Citations (3)

Summary

We haven't generated a summary for this paper yet.