2000 character limit reached
Principled Instructions Are All You Need for Questioning LLaMA-1/2, GPT-3.5/4 (2312.16171v2)
Published 26 Dec 2023 in cs.CL and cs.AI
Abstract: This paper introduces 26 guiding principles designed to streamline the process of querying and prompting LLMs. Our goal is to simplify the underlying concepts of formulating questions for various scales of LLMs, examining their abilities, and enhancing user comprehension on the behaviors of different scales of LLMs when feeding into different prompts. Extensive experiments are conducted on LLaMA-1/2 (7B, 13B and 70B), GPT-3.5/4 to verify the effectiveness of the proposed principles on instructions and prompts design. We hope that this work can provide a better guide for researchers working on the prompting of LLMs. Project page is available at https://github.com/VILA-Lab/ATLAS.
- Ask me anything: A simple strategy for prompting language models, 2022.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, abs/1810.04805, 2018.
- Training compute-optimal large language models, 2022.
- Mathprompter: Mathematical reasoning using large language models. arXiv preprint arXiv:2303.05398, 2023.
- Mistral 7b, 2023.
- Evaluating open-domain question answering in the era of large language models. arXiv preprint arXiv:2305.06984, 2023.
- Can language models learn from explanations in context? In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang, editors, Findings of the Association for Computational Linguistics: EMNLP 2022, pages 537–563, Abu Dhabi, United Arab Emirates, Dec. 2022. Association for Computational Linguistics.
- Starcoder: may the source be with you! arXiv preprint arXiv:2305.06161, 2023.
- Alpacaeval: An automatic evaluator of instruction-following models. https://github.com/tatsu-lab/alpaca_eval, 2023.
- Competition-level code generation with alphacode. Science, 378(6624):1092–1097, 2022.
- Guiding large language models via directional stimulus prompting. arXiv preprint arXiv:2302.11520, 2023.
- Gpt-4 technical report, 2023.
- Improving language understanding by generative pre-training. 2018.
- Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
- Scaling language models: Methods, analysis & insights from training gopher. CoRR, abs/2112.11446, 2021.
- Exploring the limits of transfer learning with a unified text-to-text transformer. CoRR, abs/1910.10683, 2019.
- Atlas: A llm inquiry principle benchmark. Preprint, 2024.
- Autoprompt: Eliciting knowledge from language models with automatically generated prompts, 2020.
- Gemini: A family of highly capable multimodal models, 2023.
- Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
- Llama 2: Open foundation and fine-tuned chat models, 2023.
- Chain-of-thought prompting elicits reasoning in large language models, 2023.
- A prompt pattern catalog to enhance prompt engineering with chatgpt, 2023.
- Judging llm-as-a-judge with mt-bench and chatbot arena. arXiv preprint arXiv:2306.05685, 2023.
- Least-to-most prompting enables complex reasoning in large language models, 2023.