Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multi-User Chat Assistant (MUCA): a Framework Using LLMs to Facilitate Group Conversations (2401.04883v4)

Published 10 Jan 2024 in cs.CL

Abstract: Recent advancements in LLMs have provided a new avenue for chatbot development. Most existing research, however, has primarily centered on single-user chatbots that determine "What" to answer. This paper highlights the complexity of multi-user chatbots, introducing the 3W design dimensions: "What" to say, "When" to respond, and "Who" to answer. Additionally, we proposed Multi-User Chat Assistant (MUCA), an LLM-based framework tailored for group discussions. MUCA consists of three main modules: Sub-topic Generator, Dialog Analyzer, and Conversational Strategies Arbitrator. These modules jointly determine suitable response contents, timings, and appropriate addressees. This paper further proposes an LLM-based Multi-User Simulator (MUS) to ease MUCA's optimization, enabling faster simulation of conversations between the chatbot and simulated users, and speeding up MUCA's early development. In goal-oriented conversations with a small to medium number of participants, MUCA demonstrates effectiveness in tasks like chiming in at appropriate timings, generating relevant content, and improving user engagement, as shown by case studies and user studies.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (62)
  1. Palm 2 technical report. arXiv preprint arXiv:2305.10403.
  2. Searchbots: User engagement with chatbots during collaborative search. Proceedings of the 2018 Conference on Human Information Interaction & Retrieval.
  3. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
  4. Paweł Budzianowski and Ivan Vulić. 2019. Hello, it’s gpt-2–how can i help you? towards the use of pretrained language models for task-oriented dialogue systems. arXiv preprint arXiv:1907.05774.
  5. Evaluating large language models trained on code. arXiv preprint arXiv:2107.03374.
  6. Calendar.help: Designing a workflow-based scheduling agent with humans in the loop. In Proceedings of the 2017 CHI Conference on Human Factors in Computing Systems, CHI ’17. ACM.
  7. How should the agent communicate to the group? communication strategies of a conversational agent in group chat discussions. Proceedings of the ACM on Human-Computer Interaction, 6(CSCW2):1–23.
  8. Isabella Graßl. 2019. A survey on reinforcement learning for dialogue systems. viXra.
  9. Learning from dialogue after deployment: Feed yourself, chatbot!
  10. Space-2: Tree-structured semi-supervised contrastive pre-training for task-oriented dialog understanding. arXiv preprint arXiv:2209.06638.
  11. Galaxy: A generative pre-trained model for task-oriented dialog with semi-supervised learning and explicit policy injection. In Proceedings of the AAAI conference on artificial intelligence, volume 36, pages 10749–10757.
  12. A simple language model for task-oriented dialogue. Advances in Neural Information Processing Systems, 33:20179–20191.
  13. Way off-policy batch deep reinforcement learning of implicit human preferences in dialog.
  14. Cheerbots: Chatbots toward empathy and emotionusing reinforcement learning.
  15. Bot in the bunch: Facilitating group chat discussion by improving efficiency and participation with a chatbot. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pages 1–13.
  16. Speech sentiment and customer satisfaction estimation in socialbot conversations.
  17. Offline reinforcement learning from human feedback in real-world sequence-to-sequence tasks.
  18. A survey on recent advances and challenges in reinforcement learning methods for task-oriented dialogue policy learning. Machine Intelligence Research.
  19. Prompted llms as chatbot modules for long open-domain conversation. arXiv preprint arXiv:2305.04533.
  20. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems, 33:9459–9474.
  21. Deep reinforcement learning for dialogue generation.
  22. End-to-end task-completion neural dialogue systems.
  23. Encouraging divergent thinking in large language models through multi-agent debate. arXiv preprint arXiv:2305.19118.
  24. Bing Liu and Ian Lane. 2017. Iterative policy learning in end-to-end trainable task-oriented neural dialog models.
  25. Chameleon: Plug-and-play compositional reasoning with large language models. arXiv preprint arXiv:2304.09842.
  26. A holistic approach to undesired content detection in the real world.
  27. OpenAI. 2023. Gpt-4 technical report.
  28. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744.
  29. Hybrid chat and task dialogue for more engaging hri using reinforcement learning. In 2017 26th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), pages 593–598.
  30. Language models are unsupervised multitask learners.
  31. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  32. Chinnadhurai Sankar and Sujith Ravi. 2019. Deep reinforcement learning for modeling chit-chat dialog with discrete attributes.
  33. Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100.
  34. Agenda-based user simulation for bootstrapping a POMDP dialogue system. In Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Companion Volume, Short Papers, pages 149–152, Rochester, New York. Association for Computational Linguistics.
  35. A survey of statistical user simulation techniques for reinforcement-learning of dialogue management strategies. Knowl. Eng. Rev., 21(2):97–126.
  36. Emanuel A. Schegloff. 1968. Sequencing in conversational openings. American Anthropologist, 70:1075–1095.
  37. Abigail See and Christopher Manning. 2021. Understanding and predicting user dissatisfaction in a neural generative chatbot. In Proceedings of the 22nd Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 1–12, Singapore and Online. Association for Computational Linguistics.
  38. A deep reinforcement learning chatbot.
  39. On-line active reward learning for policy optimisation in spoken dialogue systems. CoRR, abs/1605.07669.
  40. Learning from real users: Rating dialogue success with neural networks for reinforcement learning in spoken dialogue systems.
  41. Multi-task pre-training for plug-and-play task-oriented dialogue system. arXiv preprint arXiv:2109.14739.
  42. Understanding the capabilities, limitations, and societal impact of large language models.
  43. Lamda: Language models for dialog applications. arXiv preprint arXiv:2201.08239.
  44. A self learning chat-bot from user interactions and preferences. In 2020 4th International Conference on Intelligent Computing and Control Systems (ICICCS), pages 224–229.
  45. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
  46. Understanding chatbot-mediated task management. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, CHI ’18. ACM.
  47. Improving chatbot responses with context and deep seq2seq reinforcement learning.
  48. Comparing moderation strategies in group chats with multi-user chatbots. In Proceedings of the 4th Conference on Conversational User Interfaces, pages 1–4.
  49. Marilyn Walker and Steve Whittaker. 1990. Mixed initiative in dialogue: An investigation into discourse segmentation. In 28th Annual Meeting of the Association for Computational Linguistics, pages 70–78, Pittsburgh, Pennsylvania, USA. Association for Computational Linguistics.
  50. PARADISE: A framework for evaluating spoken dialogue agents. In 35th Annual Meeting of the Association for Computational Linguistics and 8th Conference of the European Chapter of the Association for Computational Linguistics, pages 271–280, Madrid, Spain. Association for Computational Linguistics.
  51. Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models.
  52. Task-oriented dialogue system as natural language generation. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2698–2703.
  53. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171.
  54. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
  55. Jason D. Williams and Geoffrey Zweig. 2016. End-to-end lstm-based dialog control optimized with supervised and reinforcement learning.
  56. Autogen: Enabling next-gen llm applications via multi-agent conversation framework. arXiv preprint arXiv:2308.08155.
  57. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564.
  58. Ubar: Towards fully end-to-end task-oriented dialog system with gpt-2. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 35, pages 14230–14238.
  59. Tree of thoughts: Deliberate problem solving with large language models. arXiv preprint arXiv:2305.10601.
  60. prompt-based statistical spoken dialog systems: A review. Proceedings of the IEEE, 101(5):1160–1179.
  61. Learning conversational systems that interleave task and non-task content.
  62. Strategy and policy learning for non-task-oriented conversational systems. In Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue, pages 404–412, Los Angeles. Association for Computational Linguistics.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Manqing Mao (1 paper)
  2. Paishun Ting (4 papers)
  3. Yijian Xiang (4 papers)
  4. Mingyang Xu (8 papers)
  5. Julia Chen (2 papers)
  6. Jianzhe Lin (15 papers)
Citations (1)