Papers
Topics
Authors
Recent
2000 character limit reached

Retrofitting Structure-aware Transformer Language Model for End Tasks (2009.07408v1)

Published 16 Sep 2020 in cs.CL

Abstract: We consider retrofitting structure-aware Transformer-based LLM for facilitating end tasks by proposing to exploit syntactic distance to encode both the phrasal constituency and dependency connection into the LLM. A middle-layer structural learning strategy is leveraged for structure integration, accomplished with main semantic task training under multi-task learning scheme. Experimental results show that the retrofitted structure-aware Transformer LLM achieves improved perplexity, meanwhile inducing accurate syntactic phrases. By performing structure-aware fine-tuning, our model achieves significant improvements for both semantic- and syntactic-dependent tasks.

Citations (43)

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.