Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

MEMEN: Multi-layer Embedding with Memory Networks for Machine Comprehension (1707.09098v1)

Published 28 Jul 2017 in cs.AI and cs.CL

Abstract: Machine comprehension(MC) style question answering is a representative problem in natural language processing. Previous methods rarely spend time on the improvement of encoding layer, especially the embedding of syntactic information and name entity of the words, which are very crucial to the quality of encoding. Moreover, existing attention methods represent each query word as a vector or use a single vector to represent the whole query sentence, neither of them can handle the proper weight of the key words in query sentence. In this paper, we introduce a novel neural network architecture called Multi-layer Embedding with Memory Network(MEMEN) for machine reading task. In the encoding layer, we employ classic skip-gram model to the syntactic and semantic information of the words to train a new kind of embedding layer. We also propose a memory network of full-orientation matching of the query and passage to catch more pivotal information. Experiments show that our model has competitive results both from the perspectives of precision and efficiency in Stanford Question Answering Dataset(SQuAD) among all published results and achieves the state-of-the-art results on TriviaQA dataset.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Boyuan Pan (30 papers)
  2. Hao Li (803 papers)
  3. Zhou Zhao (219 papers)
  4. Bin Cao (51 papers)
  5. Deng Cai (181 papers)
  6. Xiaofei He (70 papers)
Citations (84)

Summary

We haven't generated a summary for this paper yet.