Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

QASem Parsing: Text-to-text Modeling of QA-based Semantics (2205.11413v2)

Published 23 May 2022 in cs.CL

Abstract: Several recent works have suggested to represent semantic relations with questions and answers, decomposing textual information into separate interrogative natural language statements. In this paper, we consider three QA-based semantic tasks - namely, QA-SRL, QANom and QADiscourse, each targeting a certain type of predication - and propose to regard them as jointly providing a comprehensive representation of textual information. To promote this goal, we investigate how to best utilize the power of sequence-to-sequence (seq2seq) pre-trained LLMs, within the unique setup of semi-structured outputs, consisting of an unordered set of question-answer pairs. We examine different input and output linearization strategies, and assess the effect of multitask learning and of simple data augmentation techniques in the setting of imbalanced training data. Consequently, we release the first unified QASem parsing tool, practical for downstream applications who can benefit from an explicit, QA-based account of information units in a text.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Ayal Klein (7 papers)
  2. Eran Hirsch (13 papers)
  3. Ron Eliav (5 papers)
  4. Valentina Pyatkin (34 papers)
  5. Avi Caciularu (46 papers)
  6. Ido Dagan (72 papers)
Citations (11)

Summary

We haven't generated a summary for this paper yet.