Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DiagGPT: An LLM-based and Multi-agent Dialogue System with Automatic Topic Management for Flexible Task-Oriented Dialogue (2308.08043v4)

Published 15 Aug 2023 in cs.CL and cs.AI

Abstract: A significant application of LLMs, like ChatGPT, is their deployment as chat agents, which respond to human inquiries across a variety of domains. While current LLMs proficiently answer general questions, they often fall short in complex diagnostic scenarios such as legal, medical, or other specialized consultations. These scenarios typically require Task-Oriented Dialogue (TOD), where an AI chat agent must proactively pose questions and guide users toward specific goals or task completion. Previous fine-tuning models have underperformed in TOD and the full potential of conversational capability in current LLMs has not yet been fully explored. In this paper, we introduce DiagGPT (Dialogue in Diagnosis GPT), an innovative approach that extends LLMs to more TOD scenarios. In addition to guiding users to complete tasks, DiagGPT can effectively manage the status of all topics throughout the dialogue development. This feature enhances user experience and offers a more flexible interaction in TOD. Our experiments demonstrate that DiagGPT exhibits outstanding performance in conducting TOD with users, showing its potential for practical applications in various fields.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. Recent neural methods on dialogue state tracking for task-oriented dialogue systems: A survey. In Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 239–251, Singapore and Online. Association for Computational Linguistics.
  2. Task-optimized adapters for an end-to-end task-oriented dialogue system. In Findings of the Association for Computational Linguistics: ACL 2023, pages 7355–7369, Toronto, Canada. Association for Computational Linguistics.
  3. Language models are few-shot learners. In Advances in Neural Information Processing Systems, volume 33, pages 1877–1901. Curran Associates, Inc.
  4. Palm: Scaling language modeling with pathways.
  5. Schema-guided user satisfaction modeling for task-oriented dialogues. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2079–2091, Toronto, Canada. Association for Computational Linguistics.
  6. Vojtěch Hudeček and Ondřej Dušek. 2023. Are llms all you need for task-oriented dialogue?
  7. C5: Towards better conversation comprehension and contextual continuity for chatgpt.
  8. One cannot stand for everyone! leveraging multiple user simulators to train task-oriented dialogue systems. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1–21, Toronto, Canada. Association for Computational Linguistics.
  9. Rethinking the role of demonstrations: What makes in-context learning work? In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11048–11064, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  10. Multi3NLU++: A multilingual, multi-intent, multi-domain dataset for natural language understanding in task-oriented dialogue. In Findings of the Association for Computational Linguistics: ACL 2023, pages 3732–3755, Toronto, Canada. Association for Computational Linguistics.
  11. OpenAI. 2023. Gpt-4 technical report.
  12. Toolformer: Language models can teach themselves to use tools.
  13. Hugginggpt: Solving ai tasks with chatgpt and its friends in hugging face.
  14. Emergent abilities of large language models.
  15. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
  16. A network-based end-to-end trainable task-oriented dialogue system. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 438–449, Valencia, Spain. Association for Computational Linguistics.
  17. Transferable multi-domain state generator for task-oriented dialogue systems. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 808–819, Florence, Italy. Association for Computational Linguistics.
  18. An explanation of in-context learning as implicit bayesian inference. In International Conference on Learning Representations.
  19. Evalai: Towards better evaluation systems for ai agents.
  20. Ask an expert: Leveraging language models to improve strategic reasoning in goal-oriented dialogue models.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Lang Cao (19 papers)
Citations (2)