Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Span-ConveRT: Few-shot Span Extraction for Dialog with Pretrained Conversational Representations (2005.08866v2)

Published 18 May 2020 in cs.CL

Abstract: We introduce Span-ConveRT, a light-weight model for dialog slot-filling which frames the task as a turn-based span extraction task. This formulation allows for a simple integration of conversational knowledge coded in large pretrained conversational models such as ConveRT (Henderson et al., 2019). We show that leveraging such knowledge in Span-ConveRT is especially useful for few-shot learning scenarios: we report consistent gains over 1) a span extractor that trains representations from scratch in the target domain, and 2) a BERT-based span extractor. In order to inspire more work on span extraction for the slot-filling task, we also release RESTAURANTS-8K, a new challenging data set of 8,198 utterances, compiled from actual conversations in the restaurant booking domain.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Sam Coope (6 papers)
  2. Tyler Farghly (5 papers)
  3. Daniela Gerz (11 papers)
  4. Ivan Vulić (130 papers)
  5. Matthew Henderson (13 papers)
Citations (62)

Summary

We haven't generated a summary for this paper yet.