Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
92 tokens/sec
Gemini 2.5 Pro Premium
46 tokens/sec
GPT-5 Medium
19 tokens/sec
GPT-5 High Premium
32 tokens/sec
GPT-4o
87 tokens/sec
DeepSeek R1 via Azure Premium
98 tokens/sec
GPT OSS 120B via Groq Premium
435 tokens/sec
Kimi K2 via Groq Premium
207 tokens/sec
2000 character limit reached

Introducing MAPO: Momentum-Aided Gradient Descent Prompt Optimization (2410.19499v3)

Published 25 Oct 2024 in cs.CL

Abstract: Momentum-Aided Prompt Optimization (MAPO) enhances the efficiency and efficacy of prompt optimization for LLMs. Building on ProTeGi, MAPO uses positive natural language "gradients" and a momentum-based extension to refine prompts effectively. By tracking gradient history, MAPO avoids local minima and oscillations. It also utilizes beam search and an Upper Confidence Bound (UCB) algorithm for balanced candidate expansion and selection. Benchmark testing shows that MAPO achieves faster convergence time with fewer API calls and higher F1 scores than ProTeGi, proving it as a robust and scalable solution for automated prompt engineering in LLMs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (16)
  1. RLPrompt: Optimizing Discrete Text Prompts with Reinforcement Learning. arXiv.org.
  2. PromptBreeder: Self-Referential Self-Improvement via Prompt Evolution. arXiv.org.
  3. Prompt Optimization with Human Feedback. arXiv.org.
  4. Eureka: Human-Level Reward Design via Coding Large Language Models. arXiv.org.
  5. ETHOS: A Multi-Label Hate Speech Detection Dataset. Complex & Intelligent Systems, 8(6):4663–4678.
  6. OpenAI. 2022. Introducing ChatGPT. Accessed: 2024-10-19.
  7. Automatic Prompt Optimization with “Gradient Descent” and Beam Search. arXiv.org.
  8. A Systematic Survey of Prompt Engineering in Large Language Models: Techniques and Applications. arXiv.org.
  9. The Prompt Report: A Systematic Survey of Prompting Techniques. arXiv.org.
  10. AutoPrompt: Eliciting Knowledge from Language Models with Automatically Generated Prompts. arXiv.org.
  11. Automatic Prompt Augmentation and Selection with Chain-of-Thought from Labeled Data. arXiv.org.
  12. William Yang Wang. 2017. “Liar, Liar Pants on Fire”: A New Benchmark Dataset for Fake News Detection. arXiv.org.
  13. Large Language Models as Optimizers. arXiv.org.
  14. Self-Taught Optimizer (STOP): Recursively Self-Improving Code Generation. arXiv.org.
  15. TEMPERA: Test-Time Prompting via Reinforcement Learning. arXiv.org.
  16. Large Language Models are Human-Level Prompt Engineers. arXiv.org.
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube