Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash 78 tok/s
Gemini 2.5 Pro 43 tok/s Pro
GPT-5 Medium 23 tok/s
GPT-5 High 29 tok/s Pro
GPT-4o 93 tok/s
GPT OSS 120B 470 tok/s Pro
Kimi K2 183 tok/s Pro
2000 character limit reached

Fleet of Agents: Coordinated Problem Solving with Large Language Models (2405.06691v3)

Published 7 May 2024 in cs.CL, cs.AI, cs.LG, and cs.NE

Abstract: While numerous frameworks have been developed to enhance the reasoning abilities of LLMs, there is a scarcity of methods that effectively balance the trade-off between cost and quality. In this paper, we introduce Fleet of Agents (FoA), a novel and intuitive yet principled framework utilizing LLMs as agents to navigate through dynamic tree searches, employing a genetic-type particle filtering approach. FoA spawns a multitude of agents, each exploring the search space autonomously, followed by a selection phase where resampling based on a heuristic value function optimizes the balance between exploration and exploitation. This mechanism enables dynamic branching, adapting the exploration strategy based on discovered solutions. We conduct extensive experiments on three benchmark tasks, Game of 24'',Mini-Crosswords'', and WebShop'', utilizing four different LLMs,GPT-3.5'', GPT-4'',LLaMA3.2-11B'', and ``LLaMA3.2-90B''. On average across all tasks and LLMs, FoA obtains a quality improvement of ~5% while requiring only ~40% of the cost of previous SOTA methods. Notably, our analyses reveal that (1) FoA achieves the best cost-quality trade-off among all benchmarked methods and (2) FoA + LLaMA3.2-11B surpasses the Llama3.2-90B model. FoA is publicly available at https://github.com/au-clan/FoA.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. Graph of thoughts: Solving elaborate problems with large language models. In Proceedings of the AAAI Conference on Artificial Intelligence (2024), vol. 38, pp. 17682–17690.
  2. Language models are few-shot learners. In Advances in Neural Information Processing Systems (2020), H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan, and H. Lin, Eds., vol. 33, Curran Associates, Inc., pp. 1877–1901.
  3. Chase, H. Langchain. https://github.com/hwchase17/langchain, 2022.
  4. Daniel, K. Thinking, fast and slow. 2017.
  5. Evolutionary algorithms in engineering applications. Springer Science & Business Media, 2013.
  6. Holland, J. H. Adaptation in natural and artificial systems: an introductory analysis with applications to biology, control, and artificial intelligence. MIT press, 1992.
  7. Metagpt: Meta programming for multi-agent collaborative framework. CoRR abs/2308.00352 (2023).
  8. Flows: Building blocks of reasoning and collaborating ai. arXiv preprint arXiv:2308.01285 (2023).
  9. Large language models are zero-shot reasoners. In Advances in Neural Information Processing Systems (2022), S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, Eds., vol. 35, Curran Associates, Inc., pp. 22199–22213.
  10. Positioning error compensation of an industrial robot using neural networks and experimental study. Chinese Journal of Aeronautics 35, 2 (2022), 346–360. https://www.sciencedirect.com/science/article/abs/pii/S1000936121001187 ; https://www.sciencedirect.com/science/article/pii/S1000936121001187.
  11. Camel: Communicative agents for" mind" exploration of large scale language model society. arXiv preprint arXiv:2303.17760 (2023).
  12. More agents is all you need, 2024.
  13. Chameleon: Plug-and-play compositional reasoning with large language models. ArXiv abs/2304.09842 (2023).
  14. Self-refine: Iterative refinement with self-feedback. arXiv preprint arXiv:2303.17651 (2023).
  15. A hybrid genetic - particle swarm optimization algorithm for the vehicle routing problem. Expert Systems with Applications 37, 2 (2010), 1446–1455. https://dblp.uni-trier.de/db/journals/eswa/eswa37.html#MarinakisM10 ; https://www.sciencedirect.com/science/article/pii/S0957417409006460 ; https://dl.acm.org/doi/10.1016/j.eswa.2009.06.085.
  16. Explaining machine learning classifiers through diverse counterfactual explanations. In Proceedings of the 2020 conference on fairness, accountability, and transparency (2020), pp. 607–617.
  17. Nakajima, Y. Babyagi. https://github.com/yoheinakajima/babyagi, 2023.
  18. Show your work: Scratchpads for intermediate computation with language models. CoRR abs/2112.00114 (2021).
  19. Refiner: Reasoning feedback on intermediate representations. arXiv preprint arXiv:2304.01904 (2023).
  20. Richards, T. B. Autogpt. https://github.com/Significant-Gravitas/Auto-GPT, 2023.
  21. Dynamic optimization of fuzzy cognitive maps for time series forecasting. Knowledge-Based Systems 105 (2016), 29–37. https://dl.acm.org/doi/10.1016/j.knosys.2016.04.023 ; https://dblp.uni-trier.de/db/journals/kbs/kbs105.html#SalmeronF16 ; https://dx.doi.org/10.1016/j.knosys.2016.04.023 ; http://dx.doi.org/10.1016/j.knosys.2016.04.023 ; https://www.sciencedirect.com/science/article/pii/S0950705116300752.
  22. Hugginggpt: Solving ai tasks with chatgpt and its friends in huggingface. ArXiv abs/2303.17580 (2023).
  23. Self-consistency improves chain of thought reasoning in language models. arXiv preprint arXiv:2203.11171 (2022).
  24. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35 (2022), 24824–24837.
  25. The multidimensional wisdom of crowds. Advances in neural information processing systems 23 (2010).
  26. Autogen: Enabling next-gen LLM applications via multi-agent conversation framework. CoRR abs/2308.08155 (2023).
  27. Tree of thoughts: Deliberate problem solving with large language models. Advances in Neural Information Processing Systems 36 (2024).
  28. React: Synergizing reasoning and acting in language models. In The Eleventh International Conference on Learning Representations (2023).
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Ai Generate Text Spark Streamline Icon: https://streamlinehq.com

Paper Prompts

Sign up for free to create and run prompts on this paper using GPT-5.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.