Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Towards a Universal NLG for Dialogue Systems and Simulators with Future Bridging (2105.10267v2)

Published 21 May 2021 in cs.CL, cs.AI, and cs.LG

Abstract: In a dialogue system pipeline, a natural language generation (NLG) unit converts the dialogue direction and content to a corresponding natural language realization. A recent trend for dialogue systems is to first pre-train on large datasets and then fine-tune in a supervised manner using datasets annotated with application-specific features. Though novel behaviours can be learned from custom annotation, the required effort severely bounds the quantity of the training set, and the application-specific nature limits the reuse. In light of the recent success of data-driven approaches, we propose the novel future bridging NLG (FBNLG) concept for dialogue systems and simulators. The critical step is for an FBNLG to accept a future user or system utterance to bridge the present context towards. Future bridging enables self supervised training over annotation-free datasets, decoupled the training of NLG from the rest of the system. An FBNLG, pre-trained with massive datasets, is expected to apply in classical or new dialogue scenarios with minimal adaptation effort. We evaluate a prototype FBNLG to show that future bridging can be a viable approach to a universal few-shot NLG for task-oriented and chit-chat dialogues.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Philipp Ennen (3 papers)
  2. Yen-Ting Lin (117 papers)
  3. Ferdinando Insalata (6 papers)
  4. Maolin Li (10 papers)
  5. Ye Tian (191 papers)
  6. Sepehr Jalali (5 papers)
  7. Da-shan Shiu (27 papers)
  8. Ali Girayhan Ozbay (1 paper)
Citations (2)

Summary

We haven't generated a summary for this paper yet.