Authentic user advice-seeking behavior with LLMs
Investigate and characterize what users ask for and how they naturally formulate advice-seeking prompts in authentic, real-world interactions with large language models (e.g., ChatGPT, Claude, Gemini), in order to ground user-welfare safety evaluations in observed behavior rather than stated preferences or synthetic prompt constructions.
Sponsor
References
We lack understanding of what and how users naturally ask for advice in authentic interactions. Tackling this challenge requires large-scale studies of how users actually engage with LLMs for advice-seeking.
— Challenges of Evaluating LLM Safety for User Welfare
(2512.10687 - Kempermann et al., 11 Dec 2025) in Section 5 (Discussion), subsection "Behavioural realism"