Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DAG-based Long Short-Term Memory for Neural Word Segmentation (1707.00248v1)

Published 2 Jul 2017 in cs.CL

Abstract: Neural word segmentation has attracted more and more research interests for its ability to alleviate the effort of feature engineering and utilize the external resource by the pre-trained character or word embeddings. In this paper, we propose a new neural model to incorporate the word-level information for Chinese word segmentation. Unlike the previous word-based models, our model still adopts the framework of character-based sequence labeling, which has advantages on both effectiveness and efficiency at the inference stage. To utilize the word-level information, we also propose a new long short-term memory (LSTM) architecture over directed acyclic graph (DAG). Experimental results demonstrate that our model leads to better performances than the baseline models.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (4)
  1. Xinchi Chen (15 papers)
  2. Zhan Shi (84 papers)
  3. Xipeng Qiu (257 papers)
  4. Xuanjing Huang (287 papers)
Citations (10)

Summary

We haven't generated a summary for this paper yet.