Diverse Retrieval-Augmented In-Context Learning for Dialogue State Tracking (2307.01453v1)
Abstract: There has been significant interest in zero and few-shot learning for dialogue state tracking (DST) due to the high cost of collecting and annotating task-oriented dialogues. Recent work has demonstrated that in-context learning requires very little data and zero parameter updates, and even outperforms trained methods in the few-shot setting (Hu et al. 2022). We propose RefPyDST, which advances the state of the art with three advancements to in-context learning for DST. First, we formulate DST as a Python programming task, explicitly modeling language coreference as variable reference in Python. Second, since in-context learning depends highly on the context examples, we propose a method to retrieve a diverse set of relevant examples to improve performance. Finally, we introduce a novel re-weighting method during decoding that takes into account probabilities of competing surface forms, and produces a more accurate dialogue state prediction. We evaluate our approach using MultiWOZ and achieve state-of-the-art multi-domain joint-goal accuracy in zero and few-shot settings.
- Language Models are Few-Shot Learners. arXiv:2005.14165 [cs]. ArXiv: 2005.14165.
- Multiwoz - a large-scale multi-domain wizard-of-oz dataset for task-oriented dialogue modelling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP).
- Evaluating Large Language Models Trained on Code. ArXiv:2107.03374 [cs].
- Béatrice Daille. 1994. Approche mixte pour l’extraction de terminologie : statistique lexicale et filtres linguistiques.
- MultiWOZ 2.1: A Consolidated Multi-Domain Dialogue Dataset with State Corrections and State Tracking Baselines. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pages 422–428, Marseille, France. European Language Resources Association.
- Making Pre-trained Language Models Better Few-shot Learners. arXiv:2012.15723 [cs]. ArXiv: 2012.15723.
- Jade Goldstein and Jaime Carbonell. 1998. Summarization: (1) using MMR for diversity- based reranking and (2) evaluating summaries. In TIPSTER TEXT PROGRAM PHASE III: Proceedings of a Workshop held at Baltimore, Maryland, October 13-15, 1998, pages 181–195, Baltimore, Maryland, USA. Association for Computational Linguistics.
- Show, Don’t Tell: Demonstrations Outperform Descriptions for Schema-Guided Task-Oriented Dialogue. arXiv:2204.04327 [cs]. ArXiv: 2204.04327.
- Dimensionality reduction by learning an invariant mapping. In 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), volume 2, pages 1735–1742.
- MultiWOZ 2.3: A multi-domain task-oriented dialogue dataset enhanced with annotation corrections and co-reference annotation. arXiv:2010.05594 [cs]. ArXiv: 2010.05594 version: 3.
- The Curious Case of Neural Text Degeneration. ArXiv:1904.09751 [cs].
- Surface form competition: Why the highest probability answer isn’t always right. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7038–7051, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- A Simple Language Model for Task-Oriented Dialogue. In Advances in Neural Information Processing Systems, volume 33, pages 20179–20191. Curran Associates, Inc.
- In-Context Learning for Few-Shot Dialogue State Tracking. Number: arXiv:2203.08568 arXiv:2203.08568 [cs].
- Diverse Demonstrations Improve In-context Compositional Generalization. ArXiv:2212.06800 [cs].
- Zero-shot generalization in dialog state tracking through generative question answering. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 1063–1074, Online. Association for Computational Linguistics.
- Zero-shot dialogue state tracking via cross-task transfer. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7890–7900, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- Leveraging Slot Descriptions for Zero-Shot Cross-Domain Dialogue StateTracking. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5640–5648, Online. Association for Computational Linguistics.
- What makes good in-context examples for GPT-3? In Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 100–114, Dublin, Ireland and Online. Association for Computational Linguistics.
- Few-Shot Bot: Prompt-Based Learning for Dialogue Systems. arXiv:2110.08118 [cs]. ArXiv: 2110.08118.
- Rethinking the Role of Demonstrations: What Makes In-Context Learning Work? arXiv:2202.12837 [cs]. ArXiv: 2202.12837 version: 1.
- CodeGen: An Open Large Language Model for Code with Multi-Turn Program Synthesis. ArXiv:2203.13474 [cs].
- Controllable semantic parsing via retrieval augmentation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7683–7698, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
- SOLOIST: Building Task Bots at Scale with Transfer Learning and Machine Teaching. arXiv:2005.05298 [cs]. ArXiv: 2005.05298.
- Evaluating the text-to-sql capabilities of large language models. ArXiv, abs/2204.00498.
- Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics.
- Learning To Retrieve Prompts for In-Context Learning. arXiv:2112.08633 [cs]. ArXiv: 2112.08633.
- Dialogue summaries as dialogue states (DS2), template-guided summarization for few-shot dialogue state tracking. In Findings of the Association for Computational Linguistics: ACL 2022, pages 3824–3846, Dublin, Ireland. Association for Computational Linguistics.
- Richard Shin and Benjamin Van Durme. 2022. Few-Shot Semantic Parsing with Language Models Trained on Code. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5417–5425, Seattle, United States. Association for Computational Linguistics.
- MPNet: Masked and Permuted Pre-training for Language Understanding. ArXiv:2004.09297 [cs].
- Multi-Task Pre-Training for Plug-and-Play Task-Oriented Dialogue System. arXiv:2109.14739 [cs]. ArXiv: 2109.14739.
- DiSTRICT: Dialogue State Tracking with Retriever Driven In-Context Tuning. ArXiv:2212.02851 [cs].
- Benchmarking language-agnostic intent classification for virtual assistant platforms. In Proceedings of the Workshop on Multilingual Information Access (MIA), pages 69–76, Seattle, USA. Association for Computational Linguistics.
- Improving Limited Labeled Dialogue State Tracking with Self-Supervision. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 4462–4472, Online. Association for Computational Linguistics.
- Transferable Multi-Domain State Generator for Task-Oriented Dialogue Systems. ArXiv:1905.08743 [cs].
- UnifiedSKG: Unifying and Multi-Tasking Structured Knowledge Grounding with Text-to-Text Language Models. Number: arXiv:2201.05966 arXiv:2201.05966 [cs].
- MultiWOZ 2.4: A Multi-Domain Task-Oriented Dialogue Dataset with Essential Annotation Corrections to Improve State Tracking Evaluation. In Proceedings of the 23rd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 351–360, Edinburgh, UK. Association for Computational Linguistics.
- Complementary Explanations for Effective In-Context Learning. ArXiv:2211.13892 [cs].
- Description-Driven Task-Oriented Dialog Modeling. Number: arXiv:2201.08904 arXiv:2201.08904 [cs].
- Calibrate Before Use: Improving Few-Shot Performance of Language Models. arXiv:2102.09690 [cs]. ArXiv: 2102.09690.
- Brendan King (3 papers)
- Jeffrey Flanigan (18 papers)