Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GODEL: Large-Scale Pre-Training for Goal-Directed Dialog (2206.11309v1)

Published 22 Jun 2022 in cs.CL

Abstract: We introduce GODEL (Grounded Open Dialogue LLM), a large pre-trained LLM for dialog. In contrast with earlier models such as DialoGPT, GODEL leverages a new phase of grounded pre-training designed to better support adapting GODEL to a wide range of downstream dialog tasks that require information external to the current conversation (e.g., a database or document) to produce good responses. Experiments against an array of benchmarks that encompass task-oriented dialog, conversational QA, and grounded open-domain dialog show that GODEL outperforms state-of-the-art pre-trained dialog models in few-shot fine-tuning setups, in terms of both human and automatic evaluation. A novel feature of our evaluation methodology is the introduction of a notion of utility that assesses the usefulness of responses (extrinsic evaluation) in addition to their communicative features (intrinsic evaluation). We show that extrinsic evaluation offers improved inter-annotator agreement and correlation with automated metrics. Code and data processing scripts are publicly available.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Baolin Peng (72 papers)
  2. Michel Galley (50 papers)
  3. Pengcheng He (60 papers)
  4. Chris Brockett (37 papers)
  5. Lars Liden (12 papers)
  6. Elnaz Nouri (14 papers)
  7. Zhou Yu (206 papers)
  8. Bill Dolan (45 papers)
  9. Jianfeng Gao (344 papers)
Citations (67)