Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Subword ELMo (1909.08357v1)

Published 18 Sep 2019 in cs.CL

Abstract: Embedding from LLMs (ELMo) has shown to be effective for improving many NLP tasks, and ELMo takes character information to compose word representation to train LLMs.However, the character is an insufficient and unnatural linguistic unit for word representation.Thus we introduce Embedding from Subword-aware LLMs (ESuLMo) which learns word representation from subwords using unsupervised segmentation over words.We show that ESuLMo can enhance four benchmark NLP tasks more effectively than ELMo, including syntactic dependency parsing, semantic role labeling, implicit discourse relation recognition and textual entailment, which brings a meaningful improvement over ELMo.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Jiangtong Li (24 papers)
  2. Hai Zhao (227 papers)
  3. Zuchao Li (76 papers)
  4. Wei Bi (62 papers)
  5. Xiaojiang Liu (27 papers)
Citations (1)