Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Principled Instructions Are All You Need for Questioning LLaMA-1/2, GPT-3.5/4 (2312.16171v2)

Published 26 Dec 2023 in cs.CL and cs.AI

Abstract: This paper introduces 26 guiding principles designed to streamline the process of querying and prompting LLMs. Our goal is to simplify the underlying concepts of formulating questions for various scales of LLMs, examining their abilities, and enhancing user comprehension on the behaviors of different scales of LLMs when feeding into different prompts. Extensive experiments are conducted on LLaMA-1/2 (7B, 13B and 70B), GPT-3.5/4 to verify the effectiveness of the proposed principles on instructions and prompts design. We hope that this work can provide a better guide for researchers working on the prompting of LLMs. Project page is available at https://github.com/VILA-Lab/ATLAS.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (26)
  1. Ask me anything: A simple strategy for prompting language models, 2022.
  2. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
  3. BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, abs/1810.04805, 2018.
  4. Training compute-optimal large language models, 2022.
  5. Mathprompter: Mathematical reasoning using large language models. arXiv preprint arXiv:2303.05398, 2023.
  6. Mistral 7b, 2023.
  7. Evaluating open-domain question answering in the era of large language models. arXiv preprint arXiv:2305.06984, 2023.
  8. Can language models learn from explanations in context? In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang, editors, Findings of the Association for Computational Linguistics: EMNLP 2022, pages 537–563, Abu Dhabi, United Arab Emirates, Dec. 2022. Association for Computational Linguistics.
  9. Starcoder: may the source be with you! arXiv preprint arXiv:2305.06161, 2023.
  10. Alpacaeval: An automatic evaluator of instruction-following models. https://github.com/tatsu-lab/alpaca_eval, 2023.
  11. Competition-level code generation with alphacode. Science, 378(6624):1092–1097, 2022.
  12. Guiding large language models via directional stimulus prompting. arXiv preprint arXiv:2302.11520, 2023.
  13. Gpt-4 technical report, 2023.
  14. Improving language understanding by generative pre-training. 2018.
  15. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019.
  16. Scaling language models: Methods, analysis & insights from training gopher. CoRR, abs/2112.11446, 2021.
  17. Exploring the limits of transfer learning with a unified text-to-text transformer. CoRR, abs/1910.10683, 2019.
  18. Atlas: A llm inquiry principle benchmark. Preprint, 2024.
  19. Autoprompt: Eliciting knowledge from language models with automatically generated prompts, 2020.
  20. Gemini: A family of highly capable multimodal models, 2023.
  21. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971, 2023.
  22. Llama 2: Open foundation and fine-tuned chat models, 2023.
  23. Chain-of-thought prompting elicits reasoning in large language models, 2023.
  24. A prompt pattern catalog to enhance prompt engineering with chatgpt, 2023.
  25. Judging llm-as-a-judge with mt-bench and chatbot arena. arXiv preprint arXiv:2306.05685, 2023.
  26. Least-to-most prompting enables complex reasoning in large language models, 2023.
Citations (49)

Summary

  • The paper presents 26 prompt engineering guidelines that significantly enhance the clarity and precision of LLM responses.
  • Empirical evaluations using the ATLAS benchmark demonstrated quality improvements of 57.7% and accuracy gains of 67.3% across models.
  • The research lays a foundation for integrating structured prompt strategies into LLM fine-tuning and diverse automated applications.

Introduction

In the field of natural language processing, the interaction with LLMs is refined through the art of prompt engineering, the process of computer-assisted or manual crafting of instructions to extract specific responses from the models. This research explores the complexity of prompting LLMs by introducing a set of principled instructions that aim to improve the prompting process for users.

The Principles of Prompt Engineering

The paper introduces 26 principles categorized into five groups that serve to guide users in crafting prompts that would yield more precise, informative, and unbiased responses from the model. These principles encompass aspects such as prompt structure, clarity, specificity, user interaction, content style, and strategies for dealing with complex tasks and coding prompts.

An example of such a principle is to be direct and bypass polite addendums (e.g., "please," "thank you") to enhance prompt conciseness. Another encourages including the intended audience in the prompt, specifying who the response is for, such as experts or children, to tailor the LLM's output.

Empirical Results

Extensive experiments on different models, ranging from small 7B parameter models to the larger GPT-3.5 and 4, demonstrated that applying these principles led to a significant improvement in both the quality and accuracy of LLM responses. Specifically, on the ATLAS benchmark—a dataset created to evaluate principled prompt effects—the tailored prompts elevated the quality and accuracy of LLM responses by an average of 57.7% and 67.3%, respectively.

Conclusion and Future Work

The effectiveness of structured prompts in enhancing response quality from pre-trained LLMs is evident from this research. Future studies could focus on integrating these principles into regular LLM operations through fine-tuning or diversified prompting methods. This work paves the way for users to steer LLM outputs more effectively, leading to interactions with AI that are increasingly reliable and context-aware.

Youtube Logo Streamline Icon: https://streamlinehq.com
Reddit Logo Streamline Icon: https://streamlinehq.com