Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Open-Domain Frame Semantic Parsing Using Transformers (2010.10998v2)

Published 21 Oct 2020 in cs.CL and cs.AI

Abstract: Frame semantic parsing is a complex problem which includes multiple underlying subtasks. Recent approaches have employed joint learning of subtasks (such as predicate and argument detection), and multi-task learning of related tasks (such as syntactic and semantic parsing). In this paper, we explore multi-task learning of all subtasks with transformer-based models. We show that a purely generative encoder-decoder architecture handily beats the previous state of the art in FrameNet 1.7 parsing, and that a mixed decoding multi-task approach achieves even better performance. Finally, we show that the multi-task model also outperforms recent state of the art systems for PropBank SRL parsing on the CoNLL 2012 benchmark.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Aditya Kalyanpur (6 papers)
  2. Or Biran (3 papers)
  3. Tom Breloff (2 papers)
  4. Jennifer Chu-Carroll (5 papers)
  5. Ariel Diertani (2 papers)
  6. Owen Rambow (26 papers)
  7. Mark Sammons (2 papers)
Citations (16)