Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Steering Conversational Large Language Models for Long Emotional Support Conversations (2402.10453v2)

Published 16 Feb 2024 in cs.CL

Abstract: In this study, we address the challenge of enabling LLMs to consistently adhere to emotional support strategies in extended conversations. We focus on the steerability of the Llama-2 and Llama-3 suite of models, examining their ability to maintain these strategies throughout interactions. To assess this, we introduce the Strategy Relevant Attention (SRA) metric, which quantifies the model's adherence to the prompted strategy through attention maps. To facilitate our study, we create a strategy-conditioned synthetic conversational dataset derived from the ESConv dataset. We also propose various baselines informed by our proposed SRA metric to address the challenge and propose a fine-tuned model that significantly enhances the steerability of the base model in following the strategy throughout the conversation. The code and data are publicly available on our GitHub.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. Improving multi-turn emotional support dialogue generation with lookahead strategy planning. In Conference on Empirical Methods in Natural Language Processing.
  2. Qlora: Efficient finetuning of quantized llms. ArXiv, abs/2305.14314.
  3. Mahshid Hosseini and Cornelia Caragea. 2021. It takes two to empathize: One to seek and one to provide. In AAAI Conference on Artificial Intelligence.
  4. Rankgen: Improving text generation with large ranking models. ArXiv, abs/2205.09726.
  5. Lost in the middle: How language models use long contexts. ArXiv, abs/2307.03172.
  6. Towards emotional support dialog systems. ArXiv, abs/2106.01144.
  7. Lenin Medeiros and Tibor Bosse. 2018. Using crowdsourcing for the development of online emotional support agents. In Practical Applications of Agents and Multi-Agent Systems.
  8. Control globally, understand locally: A global-to-local hierarchical graph network for emotional support conversation. In International Joint Conference on Artificial Intelligence.
  9. Fado: Feedback-aware double controlling network for emotional support conversation. Knowl. Based Syst., 264:110340.
  10. The nlp task effectiveness of long-range transformers. ArXiv, abs/2202.07856.
  11. Nils Reimers and Iryna Gurevych. 2019. Sentence-bert: Sentence embeddings using siamese bert-networks. In Conference on Empirical Methods in Natural Language Processing.
  12. Towards facilitating empathic conversations in online mental health support: A reinforcement learning approach. Proceedings of the Web Conference 2021.
  13. A computational approach to understanding empathy expressed in text-based mental health support. ArXiv, abs/2009.08441.
  14. Do long-range language models actually use long-range context? ArXiv, abs/2109.09115.
  15. Llama 2: Open foundation and fine-tuned chat models. ArXiv, abs/2307.09288.
  16. Misc: A mixed strategy-aware model integrating comet for emotional support conversation. ArXiv, abs/2203.13560.
  17. From language modeling to instruction following: Understanding the behavior shift in llms after instruction tuning. ArXiv, abs/2310.00492.
  18. Transesc: Smoothing emotional support conversation via turn-level state transition. In Annual Meeting of the Association for Computational Linguistics.
  19. Building emotional support chatbots in the era of llms. ArXiv, abs/2308.11584.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Navid Madani (6 papers)
  2. Sougata Saha (13 papers)
  3. Rohini Srihari (8 papers)
Citations (2)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets