Active Preference Inference using Language Models and Probabilistic Reasoning (2312.12009v2)
Abstract: Actively inferring user preferences, for example by asking good questions, is important for any human-facing decision-making system. Active inference allows such systems to adapt and personalize themselves to nuanced individual preferences. To enable this ability for instruction-tuned LLMs, one may prompt them to ask users questions to infer their preferences, transforming the LLMs into more robust, interactive systems. However, out of the box, these models are not efficient at extracting preferences: the questions they generate are not informative, requiring a high number of user interactions and impeding the usability of the downstream system. In this work, we introduce an inference-time algorithm that helps LLMs quickly infer preferences by using more informative questions. Our algorithm uses a probabilistic model whose conditional distributions are defined by prompting an LLM, and returns questions that optimize expected entropy and expected model change. Results in a simplified interactive web shopping setting with real product items show that an LLM equipped with our entropy reduction algorithm outperforms baselines with the same underlying LLM on task performance while using fewer user interactions.
- Asking clarifying questions in open-domain information-seeking conversations. In Proceedings of the 42nd international acm sigir conference on research and development in information retrieval, pp. 475–484, 2019.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017.
- Selection-inference: Exploiting large language models for interpretable logical reasoning. arXiv preprint arXiv:2205.09712, 2022.
- Language model cascades. arXiv preprint arXiv:2207.10342, 2022.
- Kevin Ellis. Human-like few-shot learning via bayesian reasoning over natural language. In Thirty-seventh Conference on Neural Information Processing Systems, 2023.
- Paul Grice. Studies in the Way of Words. Harvard University Press, 1991.
- Eliciting human preferences with language models. arXiv preprint arXiv:2310.11589, 2023.
- Show your work: Scratchpads for intermediate computation with language models, 2021. https://arxiv.org/abs/2112.00114.
- OpenAI. Gpt-4 technical report, 2023.
- Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
- Learning to ask good questions: Ranking clarification questions using neural expected value of perfect information. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2737–2746, Melbourne, Australia, July 2018. Association for Computational Linguistics. doi: 10.18653/v1/P18-1255. URL https://aclanthology.org/P18-1255.
- Burr Settles. Active learning literature survey. 2009.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
- Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
- Webshop: Towards scalable real-world web interaction with grounded language agents. In ArXiv, 2022.
- ReAct: Synergizing reasoning and acting in language models. In International Conference on Learning Representations (ICLR), 2023.
- Interactive classification by asking informative questions. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 2664–2680, Online, July 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.acl-main.237. URL https://aclanthology.org/2020.acl-main.237.
- Generating clarifying questions for information retrieval. In Proceedings of the web conference 2020, pp. 418–428, 2020.
- Large language models as commonsense knowledge for large-scale task planning. arXiv preprint arXiv:2305.14078, 2023.
- Volodymyr Kuleshov (45 papers)
- Kevin Ellis (31 papers)
- Wasu Top Piriyakulkij (3 papers)