Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bayesian Preference Elicitation with Language Models (2403.05534v1)

Published 8 Mar 2024 in cs.CL

Abstract: Aligning AI systems to users' interests requires understanding and incorporating humans' complex values and preferences. Recently, LLMs (LMs) have been used to gather information about the preferences of human users. This preference data can be used to fine-tune or guide other LMs and/or AI systems. However, LMs have been shown to struggle with crucial aspects of preference learning: quantifying uncertainty, modeling human mental states, and asking informative questions. These challenges have been addressed in other areas of machine learning, such as Bayesian Optimal Experimental Design (BOED), which focus on designing informative queries within a well-defined feature space. But these methods, in turn, are difficult to scale and apply to real-world problems where simply identifying the relevant features can be difficult. We introduce OPEN (Optimal Preference Elicitation with Natural language) a framework that uses BOED to guide the choice of informative questions and an LM to extract features and translate abstract BOED queries into natural language questions. By combining the flexibility of LMs with the rigor of BOED, OPEN can optimize the informativity of queries while remaining adaptable to real-world domains. In user studies, we find that OPEN outperforms existing LM- and BOED-based methods for preference elicitation.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Neeraj Arora and Joel Huber. 2001. Improving Parameter Estimates and Model Prediction by Aggregate Customization in Choice Experiments. Journal of Consumer Research, 28(2):273–283. _eprint: https://academic.oup.com/jcr/article-pdf/28/2/273/17927222/28-2-273.pdf.
  2. Ralph Allan Bradley and Milton E. Terry. 1952. Rank analysis of incomplete block designs: I. the method of paired comparisons. Biometrika, 39(3/4):324–345.
  3. Language models are few-shot learners.
  4. Juan Carlos Candeal-Haro and Esteban Induráin-Eraso. 1995. A note on linear utility. Economic Theory, 6(3):519–522.
  5. Adaptive design optimization: A mutual information-based approach to model discrimination in cognitive science. Neural computation, 22:887–905.
  6. Deep Reinforcement Learning from Human Preferences. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  7. Optimal bayesian design for discriminating between models with intractable likelihoods in epidemiology. Computational Statistics & Data Analysis, 124:277–297.
  8. A tutorial on particle filtering and smoothing: Fifteen years later.
  9. Sequential bayesian experiment design for optically detected magnetic resonance of nitrogen-vacancy centers. Phys. Rev. Appl., 14:054036.
  10. Particle filters: A hands-on tutorial. Sensors, 21(2).
  11. Kto: Model alignment as prospect theoretic optimization.
  12. Variational bayesian optimal experimental design.
  13. Johannes Fürnkranz and Eyke Hüllermeier. 2012. Preference Learning, pages 2669–2672. Springer US, Boston, MA.
  14. Deep bayesian active learning with image data.
  15. Novel approach to nonlinear/non-gaussian bayesian state estimation. IEE Proc. F Radar Signal Process. UK, 140(2):107.
  16. Active ranking from pairwise comparisons and when parametric assumptions do not help. The Annals of Statistics, 47(6):3099 – 3126.
  17. Bayesian active learning for classification and preference learning.
  18. IEEE. 1984. Ieee guide for software requirements specifications. IEEE Std 830-1984, pages 1–26.
  19. Efficient experimental design with marketing research applications. Journal of Marketing Research, 31(4):545–557.
  20. Human few-shot learning of compositional instructions. In Annual Meeting of the Cognitive Science Society.
  21. David D. Lewis and Jason Catlett. 1994. Heterogeneous Uncertainty Sampling for Supervised Learning. In William W. Cohen and Haym Hirsh, editors, Machine Learning Proceedings 1994, pages 148–156. Morgan Kaufmann, San Francisco (CA).
  22. Eliciting human preferences with language models.
  23. Inferring rewards from language in context.
  24. Decision-oriented dialogue for human-ai collaboration. arXiv preprint arXiv:2305.20076.
  25. D. V. Lindley. 1956. On a measure of the information provided by an experiment. The Annals of Mathematical Statistics, 27(4):986–1005.
  26. D. V. Lindley. 1972. Bayesian Statistics. Society for Industrial and Applied Mathematics.
  27. Interactively learning preference constraints in linear bandits.
  28. Contextual multi-armed bandits. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, volume 9 of Proceedings of Machine Learning Research, pages 485–492, Chia Laguna Resort, Sardinia, Italy. PMLR.
  29. OpenAI. 2023. Gpt-4 technical report.
  30. Training language models to follow instructions with human feedback.
  31. Stefan Palan and Christian Schitter. 2017. Prolific.ac—a subject pool for online experiments. Journal of Behavioral and Experimental Finance, 17:22–27.
  32. Active preference inference using language models and probabilistic reasoning.
  33. Direct preference optimization: Your language model is secretly a reward model.
  34. Modern bayesian experimental design.
  35. Active preference-based learning of reward functions. In Robotics: Science and Systems.
  36. Burr Settles. 2010. Active learning literature survey. University of Wisconsin, Madison, 52.
  37. HAL STERN. 1990. A continuum of paired comparisons models. Biometrika, 77(2):265–273. _eprint: https://academic.oup.com/biomet/article-pdf/77/2/265/5618454/77-2-265.pdf.
  38. Task ambiguity in humans and language models. In The Eleventh International Conference on Learning Representations (ICLR).
  39. Just ask for calibration: Strategies for eliciting calibrated confidence scores from language models fine-tuned with human feedback.
  40. Probabilistic polyhedral methods for adaptive choice-based conjoint analysis: Theory and application. Marketing Science, 26(5):596–610.
  41. A bayesian approach to targeted experiment design. Bioinformatics (Oxford, England), 28:1136–42.
  42. Robust active preference elicitation.
  43. Fine-tuning language models from human preferences. ArXiv, abs/1909.08593.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Kunal Handa (8 papers)
  2. Yarin Gal (170 papers)
  3. Ellie Pavlick (66 papers)
  4. Noah Goodman (57 papers)
  5. Jacob Andreas (116 papers)
  6. Alex Tamkin (29 papers)
  7. Belinda Z. Li (21 papers)
Citations (8)