Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
107 tokens/sec
Gemini 2.5 Pro Premium
58 tokens/sec
GPT-5 Medium
29 tokens/sec
GPT-5 High Premium
25 tokens/sec
GPT-4o
101 tokens/sec
DeepSeek R1 via Azure Premium
84 tokens/sec
GPT OSS 120B via Groq Premium
478 tokens/sec
Kimi K2 via Groq Premium
213 tokens/sec
2000 character limit reached

Cue-CoT: Chain-of-thought Prompting for Responding to In-depth Dialogue Questions with LLMs (2305.11792v2)

Published 19 May 2023 in cs.CL and cs.AI

Abstract: LLMs, such as \texttt{ChatGPT}, greatly empower dialogue systems with strong language understanding and generation capabilities. However, most of the previous works prompt the LLMs to directly generate a response based on the dialogue context, overlooking the underlying linguistic cues about the user status exhibited in the context. Such in-depth dialogue scenarios are challenging for existing LLMs to figure out the user's hidden needs and respond satisfactorily through a single-step inference. To this end, we propose a novel linguistic cue-based chain-of-thoughts (\textit{Cue}-CoT), which enhances the LLMs inference with an intermediate reasoning step to find cues exhibited in the dialogue, aiming to provide a more personalized and engaging response. To evaluate the approach, we build a benchmark with in-depth dialogue questions, consisting of 6 datasets in both Chinese and English, targeting 3 major linguistic cues during the conversation: \textit{personality}, \textit{emotion}, and \textit{psychology}. We conduct extensive experiments on the proposed benchmark with 5 LLMs under both zero-shot and one-shot settings. Empirical results demonstrate our proposed \textit{Cue}-CoT method outperforms standard prompting methods in terms of both \textit{helpfulness} and \textit{acceptability} on all datasets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (60)
  1. A multitask, multilingual, multimodal evaluation of chatgpt on reasoning, hallucination, and interactivity.
  2. WASSA 2022 shared task: Predicting empathy, emotion and personality in reaction to news stories. In Proceedings of the 12th Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis, pages 214–227, Dublin, Ireland. Association for Computational Linguistics.
  3. Language models are few-shot learners.
  4. Towards robust personalized dialogue generation via order-insensitive representation regularization. In Findings of the Association for Computational Linguistics: ACL 2023, pages 7337–7345, Toronto, Canada. Association for Computational Linguistics.
  5. Controllable mixed-initiative dialogue generation through prompting.
  6. Exploring the use of large language models for reference-free text quality evaluation: A preliminary empirical study. arXiv preprint arXiv:2304.00723.
  7. Pal: Persona-augmented emotional support conversation generation.
  8. Selection-inference: Exploiting large language models for interpretable logical reasoning. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net.
  9. Prompting and evaluating large language models for proactive dialogues: Clarification, target-guided, and non-collaboration. CoRR, abs/2305.13626.
  10. A unified multi-task learning framework for multi-goal conversational recommender systems. ACM Trans. Inf. Syst., 41(3):77:1–77:25.
  11. Knowledge-enhanced mixed-initiative dialogue system for emotional support conversations. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2023, pages 4079–4095.
  12. Bert: Pre-training of deep bidirectional transformers for language understanding.
  13. Active prompting with chain-of-thought for large language models.
  14. Glm: General language model pretraining with autoregressive blank infilling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 320–335.
  15. Paul Ekman. 1971. Universals and cultural differences in facial expressions of emotion. In Nebraska symposium on motivation. University of Nebraska Press.
  16. Complexity-based prompting for multi-step reasoning.
  17. COSMIC: COmmonSense knowledge for eMotion identification in conversations. In Findings of the Association for Computational Linguistics: EMNLP 2020, pages 2470–2481, Online. Association for Computational Linguistics.
  18. Team IITP-AINLPML at WASSA 2022: Empathy detection, emotion classification and personality detection. In Proceedings of the 12th Workshop on Computational Approaches to Subjectivity, Sentiment & Social Media Analysis, pages 255–260, Dublin, Ireland. Association for Computational Linguistics.
  19. Language models as zero-shot planners: Extracting actionable knowledge for embodied agents. In International Conference on Machine Learning, pages 9118–9147. PMLR.
  20. Exploring the impact of instruction data scaling on large language models: An empirical study on real-world use cases. arXiv preprint arXiv:2303.14742.
  21. Prompted LLMs as chatbot modules for long open-domain conversation. In Findings of the Association for Computational Linguistics: ACL 2023, pages 4536–4554, Toronto, Canada. Association for Computational Linguistics.
  22. Xiaonan Li and Xipeng Qiu. 2023. Finding supporting examples for in-context learning.
  23. What makes good in-context examples for GPT-3? In Proceedings of Deep Learning Inside Out (DeeLIO 2022): The 3rd Workshop on Knowledge Extraction and Integration for Deep Learning Architectures, pages 100–114, Dublin, Ireland and Online. Association for Computational Linguistics.
  24. Towards emotional support dialog systems.
  25. Intention reasoning network for multi-domain end-to-end task-oriented dialogue. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 2273–2285, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  26. Using linguistic cues for the automatic recognition of personality in conversation and text. Journal of artificial intelligence research, 30:457–500.
  27. Active learning principles for in-context learning with large language models.
  28. Lying words: Predicting deception from linguistic styles. Personality and social psychology bulletin, 29(5):665–675.
  29. Training language models to follow instructions with human feedback.
  30. Measuring and narrowing the compositionality gap in language models. CoRR, abs/2210.03350.
  31. Towards empathetic open-domain conversation models: A new benchmark and dataset. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5370–5381, Florence, Italy. Association for Computational Linguistics.
  32. Cem: Commonsense-aware empathetic response generation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 11229–11237.
  33. Lamp: When large language models meet personalization.
  34. Personality, gender, and age in the language of social media: The open-vocabulary approach. PloS one, 8(9):e73791.
  35. A computational approach to understanding empathy expressed in text-based mental health support. In EMNLP.
  36. Automatic prompt augmentation and selection with chain-of-thought from labeled data.
  37. Can you put it all together: Evaluating conversational agents’ ability to blend skills. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 2021–2030, Online. Association for Computational Linguistics.
  38. BoB: BERT over BERT for training persona-based dialogue models from limited personalized data. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 167–177, Online. Association for Computational Linguistics.
  39. PsyQA: A Chinese dataset for generating long counseling text for mental health support. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 1489–1503, Online. Association for Computational Linguistics.
  40. Stanford alpaca: An instruction-following llama model. https://github.com/tatsu-lab/stanford_alpaca.
  41. Yla R Tausczik and James W Pennebaker. 2010. The psychological meaning of words: Liwc and computerized text analysis methods. Journal of language and social psychology, 29(1):24–54.
  42. Peter D Turney. 2002. Thumbs up or thumbs down? semantic orientation applied to unsupervised classification of reviews. arXiv preprint cs/0212032.
  43. TopicRefine: Joint topic prediction and dialogue response generation for multi-turn end-to-end dialogue system. In Proceedings of the 5th International Conference on Natural Language and Speech Processing (ICNLSP 2022), pages 19–29, Trento, Italy. Association for Computational Linguistics.
  44. Tpe: Towards better compositional reasoning over conceptual tools with multi-persona collaboration.
  45. Is chatgpt a good nlg evaluator? a preliminary study. arXiv preprint arXiv:2303.04048.
  46. Plan-and-solve prompting: Improving zero-shot chain-of-thought reasoning by large language models.
  47. Large language models are not fair evaluators. arXiv preprint arXiv:2305.17926.
  48. Self-consistency improves chain of thought reasoning in language models.
  49. Emergent abilities of large language models. arXiv preprint arXiv:2206.07682.
  50. Chain-of-thought prompting elicits reasoning in large language models.
  51. D4: a Chinese dialogue dataset for depression-diagnosis-oriented chat. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 2438–2459, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  52. Star: Bootstrapping reasoning with reasoning. In NeurIPS.
  53. Personalizing dialogue agents: I have a dog, do you have pets too? In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2204–2213, Melbourne, Australia. Association for Computational Linguistics.
  54. Automatic chain of thought prompting in large language models.
  55. Is chatgpt equipped with emotional dialogue capabilities? arXiv preprint arXiv:2304.09582.
  56. CoMAE: A multi-factor hierarchical framework for empathetic response generation. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 813–824, Online. Association for Computational Linguistics.
  57. Augesc: Dialogue augmentation with large language models for emotional support conversation.
  58. A pre-training based personalized dialogue generation model with persona-sparse data. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 9693–9700.
  59. Building emotional support chatbots in the era of llms. CoRR, abs/2308.11584.
  60. A comprehensive survey on pretrained foundation models: A history from bert to chatgpt.
Citations (39)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.