Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Large Language Model Programs (2305.05364v1)

Published 9 May 2023 in cs.LG, cs.AI, and cs.CL

Abstract: In recent years, large pre-trained LLMs have demonstrated the ability to follow instructions and perform novel tasks from a few examples. The possibility to parameterise an LLM through such in-context examples widens their capability at a much lower cost than finetuning. We extend this line of reasoning and present a method which further expands the capabilities of an LLM by embedding it within an algorithm or program. To demonstrate the benefits of this approach, we present an illustrative example of evidence-supported question-answering. We obtain a 6.4\% improvement over the chain of thought baseline through a more algorithmic approach without any finetuning. Furthermore, we highlight recent work from this perspective and discuss the advantages and disadvantages in comparison to the standard approaches.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Imanol Schlag (20 papers)
  2. Sainbayar Sukhbaatar (53 papers)
  3. Asli Celikyilmaz (81 papers)
  4. Wen-tau Yih (84 papers)
  5. Jason Weston (130 papers)
  6. Jürgen Schmidhuber (124 papers)
  7. Xian Li (116 papers)
Citations (11)