Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Active Preference Inference using Language Models and Probabilistic Reasoning (2312.12009v2)

Published 19 Dec 2023 in cs.CL, cs.AI, and cs.LG

Abstract: Actively inferring user preferences, for example by asking good questions, is important for any human-facing decision-making system. Active inference allows such systems to adapt and personalize themselves to nuanced individual preferences. To enable this ability for instruction-tuned LLMs, one may prompt them to ask users questions to infer their preferences, transforming the LLMs into more robust, interactive systems. However, out of the box, these models are not efficient at extracting preferences: the questions they generate are not informative, requiring a high number of user interactions and impeding the usability of the downstream system. In this work, we introduce an inference-time algorithm that helps LLMs quickly infer preferences by using more informative questions. Our algorithm uses a probabilistic model whose conditional distributions are defined by prompting an LLM, and returns questions that optimize expected entropy and expected model change. Results in a simplified interactive web shopping setting with real product items show that an LLM equipped with our entropy reduction algorithm outperforms baselines with the same underlying LLM on task performance while using fewer user interactions.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. Asking clarifying questions in open-domain information-seeking conversations. In Proceedings of the 42nd international acm sigir conference on research and development in information retrieval, pp.  475–484, 2019.
  2. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  3. Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017.
  4. Selection-inference: Exploiting large language models for interpretable logical reasoning. arXiv preprint arXiv:2205.09712, 2022.
  5. Language model cascades. arXiv preprint arXiv:2207.10342, 2022.
  6. Kevin Ellis. Human-like few-shot learning via bayesian reasoning over natural language. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
  7. Paul Grice. Studies in the Way of Words. Harvard University Press, 1991.
  8. Eliciting human preferences with language models. arXiv preprint arXiv:2310.11589, 2023.
  9. Show your work: Scratchpads for intermediate computation with language models, 2021. https://arxiv.org/abs/2112.00114.
  10. OpenAI. Gpt-4 technical report, 2023.
  11. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
  12. Learning to ask good questions: Ranking clarification questions using neural expected value of perfect information. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp.  2737–2746, Melbourne, Australia, July 2018. Association for Computational Linguistics. doi: 10.18653/v1/P18-1255. URL https://aclanthology.org/P18-1255.
  13. Burr Settles. Active learning literature survey. 2009.
  14. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  15. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
  16. Webshop: Towards scalable real-world web interaction with grounded language agents. In ArXiv, 2022.
  17. ReAct: Synergizing reasoning and acting in language models. In International Conference on Learning Representations (ICLR), 2023.
  18. Interactive classification by asking informative questions. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp.  2664–2680, Online, July 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.acl-main.237. URL https://aclanthology.org/2020.acl-main.237.
  19. Generating clarifying questions for information retrieval. In Proceedings of the web conference 2020, pp.  418–428, 2020.
  20. Large language models as commonsense knowledge for large-scale task planning. arXiv preprint arXiv:2305.14078, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Volodymyr Kuleshov (45 papers)
  2. Kevin Ellis (31 papers)
  3. Wasu Top Piriyakulkij (3 papers)
Citations (8)