STaR-GATE: Teaching Language Models to Ask Clarifying Questions (2403.19154v3)
Abstract: When prompting LLMs to complete a task, users often leave important aspects unsaid. While asking questions could resolve this ambiguity (GATE; Li et al., 2023), models often struggle to ask good questions. We explore a LLM's ability to self-improve (STaR; Zelikman et al., 2022) by rewarding the model for generating useful questions-a simple method we dub STaR-GATE. We generate a synthetic dataset of 25,500 unique persona-task prompts to simulate conversations between a pretrained LLM-the Questioner-and a Roleplayer whose preferences are unknown to the Questioner. By asking questions, the Questioner elicits preferences from the Roleplayer. The Questioner is iteratively finetuned on questions that increase the probability of high-quality responses to the task, which are generated by an Oracle with access to the Roleplayer's latent preferences. After two iterations of self-improvement, the Questioner asks better questions, allowing it to generate responses that are preferred over responses from the initial model on 72% of tasks. Our results indicate that teaching a LLM to ask better questions leads to better personalized responses.
- Thinking fast and slow with deep learning and tree search. Advances in neural information processing systems, 30, 2017.
- Constitutional ai: Harmlessness from ai feedback. arXiv preprint arXiv:2212.08073, 2022.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877--1901, 2020.
- Deep reinforcement learning from human preferences. Advances in neural information processing systems, 30, 2017.
- Kto: Model alignment as prospect theoretic optimization. arXiv preprint arXiv:2402.01306, 2024.
- Probabilistic model-agnostic meta-learning. Advances in neural information processing systems, 31, 2018.
- Social contract ai: Aligning ai assistants with implicit group norms. arXiv preprint arXiv:2310.17769, 2023.
- Strategic reasoning with language models. arXiv preprint arXiv:2305.19165, 2023.
- Bayesian preference elicitation with language models. arXiv preprint arXiv:2403.05534, 2024.
- Training chain-of-thought via latent-variable inference. Advances in Neural Information Processing Systems, 36, 2024.
- Zero-shot goal-directed dialogue via rl on imagined conversations. arXiv preprint arXiv:2311.05584, 2023.
- V-star: Training verifiers for self-taught reasoners. arXiv preprint arXiv:2402.06457, 2024.
- Mistral 7b. arXiv preprint arXiv:2310.06825, 2023.
- Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024.
- Chatgpt for good? on opportunities and challenges of large language models for education. Learning and individual differences, 103:102274, 2023.
- Eliciting human preferences with language models. arXiv preprint arXiv:2310.11589, 2023.
- Decision-oriented dialogue for human-ai collaboration. arXiv preprint arXiv:2305.20076, 2023.
- Gemma: Open models based on gemini research and technology. arXiv preprint arXiv:2403.08295, 2024.
- Prodigy: a profile-based dialogue generation dataset, 2023.
- OpenAI. GPT-4 Technical Report. arXiv preprint arXiv:2303.08774, 2023.
- Active preference inference using language models and probabilistic reasoning. arXiv preprint arXiv:2312.12009, 2023.
- Certified deductive reasoning with language models. 2023.
- Autoact: Automatic agent learning from scratch via self-planning. arXiv preprint arXiv:2401.05268, 2024.
- Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
- Direct preference optimization: Your language model is secretly a reward model. Advances in Neural Information Processing Systems, 36, 2024.
- Explain yourself! leveraging language models for commonsense reasoning. arXiv preprint arXiv:1906.02361, 2019.
- Grounding or guesswork? large language models are presumptive grounders. arXiv preprint arXiv:2311.09144, 2023.
- Reflexion: Language agents with verbal reinforcement learning. Advances in Neural Information Processing Systems, 36, 2024.
- Mastering chess and shogi by self-play with a general reinforcement learning algorithm. arXiv preprint arXiv:1712.01815, 2017.
- Commonsenseqa: A question answering challenge targeting commonsense knowledge. arXiv preprint arXiv:1811.00937, 2018.
- Task ambiguity in humans and language models. arXiv preprint arXiv:2212.10711, 2022.
- Large language models in medicine. Nature medicine, 29(8):1930--1940, 2023.
- Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
- Large language models are not fair evaluators. arXiv preprint arXiv:2305.17926, 2023.
- Ronald J Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine learning, 8:229--256, 1992.
- Webshop: Towards scalable real-world web interaction with grounded language agents. Advances in Neural Information Processing Systems, 35:20744--20757, 2022.
- Star: Bootstrapping reasoning with reasoning. Advances in Neural Information Processing Systems, 35:15476--15488, 2022.
- Quiet-star: Language models can teach themselves to think before speaking. arXiv preprint arXiv:2403.09629, 2024.
- In-context principle learning from mistakes. arXiv preprint arXiv:2402.05403, 2024.
- Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36, 2024.
- Chinmaya Andukuri (1 paper)
- Jan-Philipp Fränken (12 papers)
- Tobias Gerstenberg (18 papers)
- Noah D. Goodman (83 papers)