Towards Generalist Prompting for Large Language Models by Mental Models (2402.18252v1)
Abstract: LLMs have demonstrated impressive performance on many tasks. However, to achieve optimal performance, specially designed prompting methods are still needed. These methods either rely on task-specific few-shot examples that require a certain level of domain knowledge, or are designed to be simple but only perform well on a few types of tasks. In this work, we attempt to introduce the concept of generalist prompting, which operates on the design principle of achieving optimal or near-optimal performance on a wide range of tasks while eliminating the need for manual selection and customization of prompts tailored to specific problems. Furthermore, we propose MeMo (Mental Models), an innovative prompting method that is simple-designed yet effectively fulfills the criteria of generalist prompting. MeMo distills the cores of various prompting methods into individual mental models and allows LLMs to autonomously select the most suitable mental models for the problem, achieving or being near to the state-of-the-art results on diverse tasks such as STEM, logical reasoning, and commonsense reasoning in zero-shot settings. We hope that the insights presented herein will stimulate further exploration of generalist prompting methods for LLMs.
- BIG bench authors. 2023. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. Transactions on Machine Learning Research.
- Language models are few-shot learners.
- Jason P. C. Chiu and Eric Nichols. 2016. Named entity recognition with bidirectional lstm-cnns.
- Palm: Scaling language modeling with pathways.
- Training verifiers to solve math word problems.
- K.J.W. Craik. 1943. The Nature of Explanation. Cambridge University Press.
- Chatlaw: Open-source legal large language model with integrated external knowledge bases.
- Implicit chain of thought reasoning via knowledge distillation.
- Chain-of-verification reduces hallucination in large language models.
- Did Aristotle Use a Laptop? A Question Answering Benchmark with Implicit Reasoning Strategies. Transactions of the Association for Computational Linguistics (TACL).
- Folio: Natural language reasoning with first-order logic.
- Measuring massive multitask language understanding.
- Llmlingua: Compressing prompts for accelerated inference of large language models.
- Longllmlingua: Accelerating and enhancing llms in long context scenarios via prompt compression.
- Philip Johnson-Laird. 1983a. Mental Models: Towards a Cognitive Science of Language, Inference and Consciousness. Harvard University Press.
- P.N. Johnson-Laird. 1983b. Mental Models: Towards a Cognitive Science of Language, Inference, and Consciousness. Cognitive science series. Harvard University Press.
- Mental models: An interdisciplinary synthesis of theory and methods. Ecology and Society, 16(1).
- Neural machine translation in linear time.
- Peter D. Kaufman. 2005. Poor Charlie’s Almanack. Walsworth Publishing Company.
- Large language models are zero-shot reasoners.
- Neural architectures for named entity recognition.
- Competition-level code generation with alphacode. Science, 378(6624):1092–1097.
- What makes good in-context examples for gpt-3333?
- Biogpt: generative pre-trained transformer for biomedical text generation and mining. Briefings in Bioinformatics, 23(6).
- An empirical study of catastrophic forgetting in large language models during continual fine-tuning.
- Effective approaches to attention-based neural machine translation.
- Orca 2: Teaching small language models how to reason.
- Can generalist foundation models outcompete special-purpose tuning? case study in medicine.
- OpenAI. 2023. Gpt-4 technical report.
- Improving language understanding by generative pre-training.
- Language models are unsupervised multitask learners.
- Significant Gravitas. 2023. AutoGPT.
- Large language models encode clinical knowledge.
- Llama: Open and efficient foundation language models.
- Llama 2: Open foundation and fine-tuned chat models.
- Promptagent: Strategic planning with language models enables expert-level prompt optimization.
- Chain-of-thought prompting elicits reasoning in large language models.
- Wikipedia. 2023a. Charlie Munger — Wikipedia, the free encyclopedia. http://en.wikipedia.org/w/index.php?title=Charlie%20Munger&oldid=1182472377. [Online; accessed 30-October-2023].
- Wikipedia. 2023b. Mental model — Wikipedia, the free encyclopedia. http://en.wikipedia.org/w/index.php?title=Mental%20model&oldid=1173089493. [Online; accessed 29-October-2023].
- Large language models as optimizers.
- Take a step back: Evoking reasoning via abstraction in large language models.
- Haoxiang Guan (4 papers)
- Jiyan He (12 papers)
- Shuxin Zheng (32 papers)
- En-Hong Chen (2 papers)
- Weiming Zhang (135 papers)
- Nenghai Yu (173 papers)