Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Better Sentence Representation with Syntax Information (2101.03343v1)

Published 9 Jan 2021 in cs.CL

Abstract: Sentence semantic understanding is a key topic in the field of natural language processing. Recently, contextualized word representations derived from pre-trained LLMs such as ELMO and BERT have shown significant improvements for a wide range of semantic tasks, e.g. question answering, text classification and sentiment analysis. However, how to add external knowledge to further improve the semantic modeling capability of model is worth probing. In this paper, we propose a novel approach to combining syntax information with a pre-trained LLM. In order to evaluate the effect of the pre-training model, first, we introduce RNN-based and Transformer-based pre-trained LLMs; secondly, to better integrate external knowledge, such as syntactic information integrate with the pre-training model, we propose a dependency syntax expansion (DSE) model. For evaluation, we have selected two subtasks: sentence completion task and biological relation extraction task. The experimental results show that our model achieves 91.2\% accuracy, outperforming the baseline model by 37.8\% on sentence completion task. And it also gets competitive performance by 75.1\% $F_{1}$ score on relation extraction task.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Chen Yang (193 papers)
Citations (1)