2000 character limit reached
A Survey on Recent Advances in LLM-Based Multi-turn Dialogue Systems (2402.18013v1)
Published 28 Feb 2024 in cs.CL and cs.AI
Abstract: This survey provides a comprehensive review of research on multi-turn dialogue systems, with a particular focus on multi-turn dialogue systems based on LLMs. This paper aims to (a) give a summary of existing LLMs and approaches for adapting LLMs to downstream tasks; (b) elaborate recent advances in multi-turn dialogue systems, covering both LLM-based open-domain dialogue (ODD) and task-oriented dialogue (TOD) systems, along with datasets and evaluation metrics; (c) discuss some future emphasis and recent research problems arising from the development of LLMs and the increasing demands on multi-turn dialogue systems.
- Joseph Weizenbaum. Eliza—a computer program for the study of natural language communication between man and machine. Communications of the ACM, 9(1):36–45, 1966.
- Artificial paranoia. Artificial intelligence, 2(1):1–25, 1971.
- David Goddeau et al. A form-based dialogue manager for spoken language applications. In Proceeding of Fourth International Conference on Spoken Language Processing. ICSLP’96, volume 2, pages 701–704. IEEE, 1996.
- Yu Wu et al. Sequential matching network: A new architecture for multi-turn response selection in retrieval-based chatbots. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 496–505, Vancouver, Canada, July 2017. Association for Computational Linguistics.
- Towards end-to-end learning for dialog state tracking and management using deep reinforcement learning. In Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 1–10, Los Angeles, September 2016. Association for Computational Linguistics.
- Wentao Ma et al. TripleNet: Triple attention network for multi-turn response selection in retrieval-based chatbots. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 737–746, Hong Kong, China, November 2019. Association for Computational Linguistics.
- Iulian Serban et al. Building end-to-end dialogue systems using generative hierarchical neural network models. In Proceedings of the AAAI conference on artificial intelligence, volume 30, 2016.
- Wanwei He et al. Amalgamating knowledge from two teachers for task-oriented dialogue system with adversarial training. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3498–3507, 2020.
- Lisong Qiu et al. Are training samples correlated? learning to generate dialogue responses with multiple references. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 3826–3835, 2019.
- Dialogue system: A brief review. arXiv preprint arXiv:1306.4134, 2013.
- Hongshen Chen et al. A survey on dialogue systems: Recent advances and new frontiers. Acm Sigkdd Explorations Newsletter, 19(2):25–35, 2017.
- Jinjie Ni et al. Recent advances in deep learning based dialogue systems: A systematic survey. Artificial intelligence review, 56(4):3055–3155, 2023.
- Libo et al. Qin. End-to-end task-oriented dialogue: A survey of tasks, methods, and future directions. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 5925–5941, Singapore, December 2023. Association for Computational Linguistics.
- Jared Kaplan et al. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361, 2020.
- Jason Wei et al. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682, 2022.
- Xipeng Qiu et al. Pre-trained models for natural language processing: A survey. Science China Technological Sciences, 63(10):1872–1897, 2020.
- Ashish Vaswani et al. Attention is all you need. Advances in neural information processing systems, 30, 2017.
- Colin Raffel et al. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485–5551, 2020.
- Peter J. Liu et al. Generating wikipedia by summarizing long sequences. arXiv preprint arXiv:1801.10198, 2018.
- Alec Radford et al. Improving language understanding by generative pre-training. 2018.
- Rami Al-Rfou et al. Character-level language modeling with deeper self-attention. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 3159–3166, 2019.
- Junjie Ye et al. A comprehensive capability analysis of gpt-3 and gpt-3.5 series models. arXiv preprint arXiv:2303.10420, 2023.
- Alec Radford et al. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
- Bryan McCann et al. The natural language decathlon: Multitask learning as question answering. arXiv preprint arXiv:1806.08730, 2018.
- Tom Brown et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Long Ouyang et al. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
- Wayne Zhao Xin et al. A survey of large language models. arXiv preprint arXiv:2303.18223, 2023.
- OpenAI. Introducing chatgpt, November 2022.
- Samantha Lock. What is ai chatbot phenomenon chatgpt and could it replace humans. The Guardian, 5, 2022.
- OpenAI. Gpt-4 technical report, 2023.
- Samantha Lock. What is ai chatbot phenomenon chatgpt and could it replace humans? The Guardian.
- Jon Gertner. Wikipedia’s moment of truth. The New York Times Magazine.
- https://www.openai.com, 2023. Retrieved October 16, 2023.
- Kevin Roose. The new chatgpt can ’see’ and ’talk.’ here’s what it’s like. The New York Times, September 27 2023. Retrieved October 16, 2023.
- OpenAI. Introducing gpts, November 2023.
- Hugo Touvron et al. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
- Meta AI. Introducing llama: A foundational, 65-billion-parameter large language model. February 2023. URL-of-the-Article (Accessed: February 28, 2024).
- Jordan Hoffmann et al. Training compute-optimal large language models. arXiv preprint arXiv:2203.15556, 2022.
- Jack W. Rae et al. Scaling language models: Methods, analysis & insights from training gopher. arXiv preprint arXiv:2112.11446, 2021.
- Meta. Meta and microsoft introduce the next generation of llama. July 2023. Retrieved 21 July 2023, from https://about.fb.com/news/2023/07/llama-2/.
- Hugo Touvron et al. Llama 2: Open foundation and fine-tuned chat models, 2023.
- Baptiste Rozière et al. Code llama: Open foundation models for code, 2023.
- Zhengxiao Du et al. Glm: General language model pretraining with autoregressive blank infilling. arXiv preprint arXiv:2103.10360, 2021.
- Aohan Zeng et al. Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414, 2022.
- THUDM. Chatglm2-6b. https://github.com/thudm/chatglm2-6b, 2022.
- THUDM. Chatglm3. https://github.com/THUDM/ChatGLM3, 2023.
- Jacob Devlin et al. BERT: Pre-training of deep bidirectional transformers for language understanding. pages 4171–4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics.
- Google AI Blog. Open sourcing bert: State-of-the-art pre-training for natural language processing. November 2018. Retrieved November 27, 2019.
- Yinhan Liu et al. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
- Li Dong et al. Unified language model pre-training for natural language understanding and generation. Advances in neural information processing systems, 32, 2019.
- Mike Lewis et al. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. arXiv preprint arXiv:1910.13461, 2019.
- Neil Houlsby et al. Parameter-efficient transfer learning for nlp. In International Conference on Machine Learning, pages 2790–2799. PMLR, 2019.
- Jonas Pfeiffer et al. AdapterHub: A framework for adapting transformers. pages 46–54, Online, October 2020. Association for Computational Linguistics.
- Xiao Liu et al. Gpt understands, too. AI Open, 2023.
- Edward Hu et al. Lora: Low-rank adaptation of large language models. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4395–4409, 2021.
- Tim Dettmers et al. Qlora: Efficient finetuning of quantized llms. arXiv preprint arXiv:2305.14314, 2023.
- Jason Wei et al. Finetuned language models are zero-shot learners. arXiv preprint arXiv:2109.01652, 2021.
- Tianyi Zhang et al. Revisiting few-sample bert fine-tuning. arXiv preprint arXiv:2006.05987, 2020.
- On the stability of fine-tuning bert: Misconceptions, explanations, and strong baselines. arXiv preprint arXiv:2006.04884, 2021.
- Armen Aghajanyan et al. Better fine-tuning by reducing representational collapse. arXiv preprint arXiv:2008.03156, 2020.
- Haoming Jiang et al. SMART: Robust and efficient fine-tuning for pre-trained natural language models through principled regularized optimization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2177–2190, Online, July 2020. Association for Computational Linguistics.
- Chen Zhu et al. Freelb: Enhanced adversarial training for natural language understanding. arXiv preprint arXiv:1909.11764, 2019.
- Exploiting cloze-questions for few-shot text classification and natural language inference. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 255–269, Online, April 2021. Association for Computational Linguistics.
- Making pre-trained language models better few-shot learners. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 3816–3830, Online, August 2021. Association for Computational Linguistics.
- Xu Han et al. Ptr: Prompt tuning with rules for text classification. AI Open, 3:182–192, 2022.
- Shengding Hu et al. Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2225–2240, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Prefix-tuning: Optimizing continuous prompts for generation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 4582–4597, Online, August 2021. Association for Computational Linguistics.
- The power of scale for parameter-efficient prompt tuning. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 3045–3059, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics.
- Xiao Liu et al. P-tuning v2: Prompt tuning can be comparable to fine-tuning universally across scales and tasks. arXiv preprint arXiv:2110.07602, 2021.
- Learning how to ask: Querying LMs with mixtures of soft prompts. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5203–5212, Online, June 2021. Association for Computational Linguistics.
- Jiachang Liu et al. What makes good in-context examples for GPT-3? In Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 100–114, Dublin, Ireland and Online, May 2022. Association for Computational Linguistics.
- Yao Lu et al. Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8086–8098, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Mingda Chen et al. Improving in-context few-shot learning via self-supervised training. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 3558–3573, Seattle, United States, July 2022. Association for Computational Linguistics.
- Yanda Chen et al. Meta-learning via language model in-context tuning. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 719–730, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Sewon Min et al. MetaICL: Learning to learn in context. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 2791–2809, Seattle, United States, July 2022. Association for Computational Linguistics.
- Jason Wei et al. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
- Takeshi Kojima et al. Large language models are zero-shot reasoners. Advances in neural information processing systems, 35:22199–22213, 2022.
- Denny Zhou et al. Least-to-most prompting enables complex reasoning in large language models. arXiv preprint arXiv:2205.10625, 2022.
- Zhuosheng Zhang et al. Automatic chain of thought prompting in large language models. arXiv preprint arXiv:2210.03493, 2022.
- Automatic prompt augmentation and selection with chain-of-thought from labeled data. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 12113–12139, Singapore, December 2023. Association for Computational Linguistics.
- Zhuosheng Zhang et al. Multimodal chain-of-thought reasoning in language models. arXiv preprint arXiv:2302.00923, 2023.
- Haoyang Huang et al. Not all languages are created equal in LLMs: Improving multilingual capability by cross-lingual-thought prompting. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 12365–12394, Singapore, December 2023. Association for Computational Linguistics.
- Li Deng et al. Use of kernel deep convex networks and end-to-end learning for spoken language understanding. In 2012 IEEE Spoken Language Technology Workshop (SLT), pages 210–215. IEEE, 2012.
- Gokhan Tur et al. Towards deeper understanding: Deep convex networks for semantic utterance classification. In 2012 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 5045–5048. IEEE, 2012.
- Daniele Comi et al. Zero-shot-bert-adapters: a zero-shot pipeline for unknown intent detection. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 650–663, 2023.
- Soham Parikh et al. Exploring zero and few-shot techniques for intent classification. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 5: Industry Track), pages 744–751, Toronto, Canada, July 2023. Association for Computational Linguistics.
- Yen-Ting Lin et al. Selective in-context data augmentation for intent detection using pointwise V-information. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, pages 1463–1476, Dubrovnik, Croatia, May 2023. Association for Computational Linguistics.
- Samuel Coope et al. Span-ConveRT: Few-shot span extraction for dialog with pretrained conversational representations. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 107–121, Online, July 2020. Association for Computational Linguistics.
- Linguistically-enriched and context-awarezero-shot slot filling. In Proceedings of the Web Conference 2021, pages 3279–3290, 2021.
- Bert for joint intent classification and slot filling. arxiv. arXiv preprint arXiv:1902.10909, 2019.
- Hoang Nguyen et al. CoF-CoT: Enhancing large language models with coarse-to-fine chain-of-thought prompting for multi-domain NLU tasks. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 12109–12119, Singapore, December 2023. Association for Computational Linguistics.
- Scalable neural dialogue state tracking. In 2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU), pages 830–837. IEEE, 2019.
- Global-locally self-attentive encoder for dialogue state tracking. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1458–1467, 2018.
- SUMBT: Slot-utterance matching for universal and scalable belief tracking. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5478–5483, Florence, Italy, July 2019. Association for Computational Linguistics.
- Slot attention with value normalization for multi-domain dialogue state tracking. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3019–3028, 2020.
- Zhaojiang Lin et al. MinTL: Minimalist transfer learning for task-oriented dialogue systems. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3391–3405, Online, November 2020. Association for Computational Linguistics.
- Yushi Hu et al. In-context learning for few-shot dialogue state tracking. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 2627–2643, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics.
- Yujie Feng et al. Towards LLM-driven dialogue state tracking. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 739–755, Singapore, December 2023. Association for Computational Linguistics.
- Multi-agent task-oriented dialog policy learning with role-aware reward decomposition. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 625–638, Online, July 2020. Association for Computational Linguistics.
- Sihan Wang et al. Task-completion dialogue policy learning via monte carlo tree search with dueling network. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3461–3471, 2020.
- Learning Dialog Policies from Weak Demonstrations. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 1394–1405, Online, July 2020. Association for Computational Linguistics.
- Wenhu Chen et al. Semantically conditioned dialog response generation via hierarchical disentangled self-attention. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 3696–3709, Florence, Italy, July 2019. Association for Computational Linguistics.
- YunHao Li et al. Retrieve & memorize: Dialog policy learning with multi-action memory. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 447–459, Online, August 2021. Association for Computational Linguistics.
- Hello, it’s GPT-2 - how can I help you? towards the use of pretrained language models for task-oriented dialogue systems. In Proceedings of the 3rd Workshop on Neural Generation and Translation, pages 15–22, Hong Kong, November 2019. Association for Computational Linguistics.
- Ehsan Hosseini-Asl et al. A simple language model for task-oriented dialogue. Advances in Neural Information Processing Systems, 33:20179–20191, 2020.
- Gpt-critic: Offline reinforcement learning for end-to-end task-oriented dialogue systems. In 10th International Conference on Learning Representations, ICLR 2022. International Conference on Learning Representations, ICLR, 2022.
- [CASPI] causal-aware safe policy improvement for task-oriented dialogue. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 92–102, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Wanwei He et al. Galaxy: A generative pre-trained model for task-oriented dialog with semi-supervised learning and explicit policy injection. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pages 10749–10757, 2022.
- Prompt-based Monte-Carlo tree search for goal-oriented dialogue policy planning. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 7101–7125, Singapore, December 2023. Association for Computational Linguistics.
- Building applied natural language generation systems. Natural Language Engineering, 3(1):57–87, 1997.
- Tsung-Hsien Wen et al. Stochastic language generation in dialogue using recurrent neural networks with convolutional sentence reranking. In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 275–284, Prague, Czech Republic, September 2015. Association for Computational Linguistics.
- Tsung-Hsien Wen et al. Semantically conditioned LSTM-based natural language generation for spoken dialogue systems. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pages 1711–1721, Lisbon, Portugal, September 2015. Association for Computational Linguistics.
- Context-aware natural language generation for spoken dialogue systems. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pages 2032–2041, 2016.
- Baolin Peng et al. Few-shot natural language generation for task-oriented dialog. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 172–182, Online, November 2020. Association for Computational Linguistics.
- Zhiyu Chen et al. Few-shot NLG with pre-trained language model. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 183–190, Online, July 2020. Association for Computational Linguistics.
- Fluent response generation for conversational question answering. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 191–207, Online, July 2020. Association for Computational Linguistics.
- Jing Qian et al. Controllable natural language generation with contrastive prefixes. In Findings of the Association for Computational Linguistics: ACL 2022, pages 2912–2924, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Baolin Peng et al. Soloist: Building task bots at scale with transfer learning and machine teaching. Transactions of the Association for Computational Linguistics, 9:807–824, 2021.
- Ubar: Towards fully end-to-end task-oriented dialog system with gpt-2. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 14230–14238, 2021.
- Yixuan Su et al. Multi-task pre-training for plug-and-play task-oriented dialogue system. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4661–4676, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Wanwei He et al. SPACE-2: Tree-structured semi-supervised contrastive pre-training for task-oriented dialog understanding. In Proceedings of the 29th International Conference on Computational Linguistics, pages 553–569, Gyeongju, Republic of Korea, October 2022. International Committee on Computational Linguistics.
- Wanwei He et al. Unified dialog model pre-training for task-oriented dialog understanding and generation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 187–200, 2022.
- Vladimir Karpukhin et al. Dense passage retrieval for open-domain question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 6769–6781, Online, November 2020. Association for Computational Linguistics.
- Chunyuan Yuan et al. Multi-hop selector network for multi-turn response selection in retrieval-based chatbots. In Proceedings of the 2019 conference on empirical methods in natural language processing and the 9th international joint conference on natural language processing (EMNLP-IJCNLP), pages 111–120, 2019.
- Chongyang Tao et al. One time of interaction may not be enough: Go deep with an interaction-over-interaction network for response selection in dialogues. In Proceedings of the 57th annual meeting of the association for computational linguistics, pages 1–11, 2019.
- Xinchao Xu et al. Long time no see! open-domain conversation with long-term persona memory. In Findings of the Association for Computational Linguistics: ACL 2022, pages 2639–2650, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Andrea Madotto et al. Personalizing dialogue agents via meta-learning. In Proceedings of the 57th annual meeting of the association for computational linguistics, pages 5454–5459, 2019.
- Ruijun Chen et al. Learning to memorize entailment and discourse relations for persona-consistent dialogues. arXiv preprint arXiv:2301.04871, 2023.
- Juntao Li et al. Dialogue history matters! personalized response selection in multi-turn retrieval-based chatbots. ACM Transactions on Information Systems (TOIS), 39(4):1–25, 2021.
- Zhengyi Ma et al. One chatbot per person: Creating personalized chatbots based on implicit user profiles. In Proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval, pages 555–564, 2021.
- Hanxun Zhong et al. Less is more: Learning to refine dialogue history for personalized dialogue generation. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 5808–5820, Seattle, United States, July 2022. Association for Computational Linguistics.
- Haoyu Song et al. Generate, delete and rewrite: A three-stage framework for improving persona consistency of dialogue generation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 5821–5831, Online, July 2020. Association for Computational Linguistics.
- Yihong Tang et al. Enhancing personalized dialogue generation with contrastive latent variables: Combining sparse and dense persona. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5456–5468, Toronto, Canada, July 2023. Association for Computational Linguistics.
- Sebastian Borgeaud et al. Improving language models by retrieving from trillions of tokens. In International conference on machine learning, pages 2206–2240. PMLR, 2022.
- Leveraging passage retrieval with generative models for open domain question answering. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, pages 874–880, Online, April 2021. Association for Computational Linguistics.
- Leonard Adolphs et al. Reason first, then respond: Modular generation for knowledge-infused dialogue. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 7112–7132, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics.
- Devendra Singh et al. End-to-end training of multi-document reader and retriever for open-domain question answering. Advances in Neural Information Processing Systems, 34:25968–25981, 2021.
- Latent retrieval for weakly supervised open domain question answering. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 6086–6096, Florence, Italy, July 2019. Association for Computational Linguistics.
- Internet-augmented dialogue generation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8460–8478, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Learning end-to-end goal-oriented dialog. arXiv preprint arXiv:1605.07683, 2016.
- Matthew Henderson et al. Efficient natural language response suggestion for smart reply. arXiv preprint arXiv:1705.00652, 2017.
- A neural conversational model. arXiv preprint arXiv:1506.05869, 2015.
- Sequence to sequence learning with neural networks. Advances in neural information processing systems, 27, 2014.
- Neural responding machine for short-text conversation. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 1577–1586, Beijing, China, July 2015. Association for Computational Linguistics.
- Xueliang Zhao et al. Knowledge-grounded dialogue generation with pre-trained language models. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 3377–3390, Online, November 2020. Association for Computational Linguistics.
- Alessandro Sordoni et al. A neural network approach to context-sensitive generation of conversational responses. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 196–205, Denver, Colorado, May–June 2015. Association for Computational Linguistics.
- Zhao Yan et al. Docchat: An information retrieval approach for chatbot engines using unstructured documents. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 516–525, 2016.
- Beyond goldfish memory: Long-term open-domain conversation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5180–5197, Dublin, Ireland, May 2022. Association for Computational Linguistics.
- Word-based dialog state tracking with recurrent neural networks. In Proceedings of the 15th annual meeting of the special interest group on discourse and dialogue (SIGDIAL), pages 292–299, 2014.
- Chien-Sheng Wu et al. Transferable multi-domain state generator for task-oriented dialogue systems. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 808–819, Florence, Italy, July 2019. Association for Computational Linguistics.
- Abhinav Rastogi et al. Towards scalable multi-domain conversational agents: The schema-guided dialogue dataset. In Proceedings of the AAAI conference on artificial intelligence, volume 34, pages 8689–8696, 2020.
- Kishore Papineni et al. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pages 311–318, 2002.
- Jiwei et al. Li. A diversity-promoting objective function for neural conversation models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 110–119, San Diego, California, June 2016. Association for Computational Linguistics.
- Paweł Budzianowski et al. MultiWOZ - a large-scale multi-domain Wizard-of-Oz dataset for task-oriented dialogue modelling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 5016–5026, Brussels, Belgium, October-November 2018. Association for Computational Linguistics.
- Mihail Eric et al. MultiWOZ 2.1: A consolidated multi-domain dialogue dataset with state corrections and state tracking baselines. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pages 422–428, Marseille, France, May 2020. European Language Resources Association.
- Xiaoxue Zang et al. MultiWOZ 2.2 : A dialogue dataset with additional annotation corrections and state tracking baselines. In Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI, pages 109–117, Online, July 2020. Association for Computational Linguistics.
- Ting Han et al. Multiwoz 2.3: A multi-domain task-oriented dialogue dataset enhanced with annotation corrections and co-reference annotation. In Natural Language Processing and Chinese Computing: 10th CCF International Conference, NLPCC 2021, Qingdao, China, October 13–17, 2021, Proceedings, Part II 10, pages 206–218. Springer, 2021.
- MultiWOZ 2.4: A multi-domain task-oriented dialogue dataset with essential annotation corrections to improve state tracking evaluation. In Proceedings of the 23rd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 351–360, Edinburgh, UK, September 2022. Association for Computational Linguistics.
- Jun Quan et al. RiSAWOZ: A large-scale multi-domain Wizard-of-Oz dataset with rich semantic annotations for task-oriented dialogue modeling. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 930–940, Online, November 2020. Association for Computational Linguistics.
- Qi Zhu et al. Crosswoz: A large-scale chinese cross-domain task-oriented dialogue dataset. Transactions of the Association for Computational Linguistics, 8:281–295, 2020.
- Xuewei Wang et al. Persuasion for good: Towards a personalized persuasive dialogue system for social good. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5635–5649, Florence, Italy, July 2019. Association for Computational Linguistics.
- Nikola Mrkšić et al. Neural belief tracker: Data-driven dialogue state tracking. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1777–1788, Vancouver, Canada, July 2017. Association for Computational Linguistics.
- Tsung-Hsien Wen et al. A network-based end-to-end trainable task-oriented dialogue system. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 438–449, Valencia, Spain, April 2017. Association for Computational Linguistics.
- Mihail et al. Eric. Key-value retrieval networks for task-oriented dialogue. In Proceedings of the 18th Annual SIGdial Meeting on Discourse and Dialogue, pages 37–49, Saarbrücken, Germany, August 2017. Association for Computational Linguistics.
- Saizheng Zhang et al. Personalizing dialogue agents: I have a dog, do you have pets too? In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2204–2213, Melbourne, Australia, July 2018. Association for Computational Linguistics.
- Jiazhan Feng et al. MMDialog: A large-scale multi-turn dialogue dataset towards multi-modal open-domain conversation. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7348–7363, July 2023.
- Yanran Li et al. DailyDialog: A manually labelled multi-turn dialogue dataset. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 986–995, Taipei, Taiwan, November 2017. Asian Federation of Natural Language Processing.
- Hongjin Qian et al. Pchatbot: a large-scale dataset for personalized chatbot. In Proceedings of the 44th international ACM SIGIR conference on research and development in information retrieval, pages 2470–2477, 2021.
- Yinhe Zheng et al. Personalized dialogue generation with diversified traits. arXiv preprint arXiv:1901.09672, 2019.
- Zihao Yi (3 papers)
- Jiarui Ouyang (2 papers)
- Yuwen Liu (3 papers)
- Tianhao Liao (1 paper)
- Zhe Xu (199 papers)
- Ying Shen (76 papers)