Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
110 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Beyond Natural Language: LLMs Leveraging Alternative Formats for Enhanced Reasoning and Communication (2402.18439v3)

Published 28 Feb 2024 in cs.CL and cs.AI

Abstract: Natural language (NL) has long been the predominant format for human cognition and communication, and by extension, has been similarly pivotal in the development and application of LLMs. Yet, besides NL, LLMs have seen various non-NL formats during pre-training, such as code and logical expression. NL's status as the optimal format for LLMs, particularly in single-LLM reasoning and multi-agent communication, has not been thoroughly examined. In this work, we challenge the default use of NL by exploring the utility of non-NL formats in these contexts. We show that allowing LLMs to autonomously select the most suitable format before reasoning or communicating leads to a 3.3 to 5.7\% improvement in reasoning efficiency for different LLMs, and up to a 72.7\% reduction in token usage in multi-agent communication, all while maintaining communicative effectiveness. Our comprehensive analysis further reveals that LLMs can devise a format from limited task instructions and that the devised format is effectively transferable across different LLMs. Intriguingly, the structured communication format decided by LLMs exhibits notable parallels with established agent communication languages, suggesting a natural evolution towards efficient, structured communication in agent communication. Our code is released at \url{https://github.com/thunlp/AutoForm}.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (38)
  1. Anthropic. 2023. Introducing claude 2.1.
  2. Emergent autonomous scientific research capabilities of large language models. CoRR, abs/2304.05332.
  3. Chateval: Towards better llm-based evaluators through multi-agent debate. CoRR, abs/2308.07201.
  4. Agentverse: Facilitating multi-agent collaboration and exploring emergent behaviors in agents. CoRR, abs/2308.10848.
  5. Program of thoughts prompting: Disentangling computation from reasoning for numerical reasoning tasks. CoRR, abs/2211.12588.
  6. Noam Chomsky. 2006. Language and mind. Cambridge University Press.
  7. Improving factuality and reasoning in language models through multiagent debate. CoRR, abs/2305.14325.
  8. KQML as an agent communication language. In Proceedings of the Third International Conference on Information and Knowledge Management (CIKM’94), Gaithersburg, Maryland, USA, November 29 - December 2, 1994, pages 456–463. ACM.
  9. FIPA. 2001. FIPA ACL Message Structure Specification. FIPA.
  10. Jerry A Fodor. 1975. The language of thought, volume 5. Harvard university press.
  11. PAL: program-aided language models. In International Conference on Machine Learning, ICML 2023, 23-29 July 2023, Honolulu, Hawaii, USA, volume 202 of Proceedings of Machine Learning Research, pages 10764–10799. PMLR.
  12. Gemini: a family of highly capable multimodal models. arXiv preprint arXiv:2312.11805.
  13. Metagpt: Meta programming for multi-agent collaborative framework. CoRR, abs/2308.00352.
  14. The narrativeqa reading comprehension challenge. Trans. Assoc. Comput. Linguistics, 6:317–328.
  15. Large language models are zero-shot reasoners. In NeurIPS.
  16. George Lakoff. 2008. Women, fire, and dangerous things: What categories reveal about the mind. University of Chicago press.
  17. CAMEL: communicative agents for "mind" exploration of large scale language model society. CoRR, abs/2303.17760.
  18. Program induction by rationale generation: Learning to solve and explain algebraic word problems. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Vancouver, Canada, July 30 - August 4, Volume 1: Long Papers, pages 158–167. Association for Computational Linguistics.
  19. Plan, verify and switch: Integrated reasoning with diverse x-of-thoughts. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023, pages 2807–2822. Association for Computational Linguistics.
  20. GAIA: a benchmark for general AI assistants. CoRR, abs/2311.12983.
  21. OpenAI. 2023a. Assistants api.
  22. OpenAI. 2023b. GPT-4 technical report. CoRR, abs/2303.08774.
  23. Anton Osika. 2023. gpt-engineer.
  24. Generative agents: Interactive simulacra of human behavior. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology, UIST 2023, San Francisco, CA, USA, 29 October 2023- 1 November 2023, pages 2:1–2:22. ACM.
  25. Let models speak ciphers: Multiagent debate through embeddings. CoRR, abs/2310.06272.
  26. Steven Pinker. 2003. The language instinct: How the mind creates language. Penguin uK.
  27. Communicative agents for software development. CoRR, abs/2307.07924.
  28. Reflexion: an autonomous agent with dynamic memory and self-reflection. CoRR, abs/2303.11366.
  29. Significant Gravitas. AutoGPT.
  30. Beyond the imitation game: Quantifying and extrapolating the capabilities of language models. CoRR, abs/2206.04615.
  31. Chain-of-thought prompting elicits reasoning in large language models. In NeurIPS.
  32. Constructing datasets for multi-hop reading comprehension across documents. Trans. Assoc. Comput. Linguistics, 6:287–302.
  33. Benjamin Lee Whorf. 2012. Language, thought, and reality: Selected writings of Benjamin Lee Whorf. MIT press.
  34. Autogen: Enabling next-gen LLM applications via multi-agent conversation framework. CoRR, abs/2308.08155.
  35. Hotpotqa: A dataset for diverse, explainable multi-hop question answering. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31 - November 4, 2018, pages 2369–2380. Association for Computational Linguistics.
  36. Tree of thoughts: Deliberate problem solving with large language models. CoRR, abs/2305.10601.
  37. React: Synergizing reasoning and acting in language models. In The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net.
  38. Webarena: A realistic web environment for building autonomous agents. CoRR, abs/2307.13854.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Weize Chen (34 papers)
  2. Chenfei Yuan (5 papers)
  3. Jiarui Yuan (5 papers)
  4. Yusheng Su (21 papers)
  5. Chen Qian (226 papers)
  6. Cheng Yang (168 papers)
  7. Ruobing Xie (97 papers)
  8. Zhiyuan Liu (433 papers)
  9. Maosong Sun (337 papers)
Citations (8)

Summary

"Beyond Natural Language: LLMs Leveraging Alternative Formats for Enhanced Reasoning and Communication" investigates the application of non-natural language formats in enhancing the reasoning and communicative efficiency of LLMs. This paper questions the conventional belief that natural language (NL) is the optimal format for cognitive processes in LLMs, particularly in reasoning and agent communication contexts.

Key findings and methodologies include:

  1. Alternative Formats: The paper explores the utilization of non-NL formats like code, logical expressions, and structured data for improving LLM reasoning and communication. The authors argue that these formats can lead to improvements in computational efficiency and clarity of communication among agents.
  2. Autonomous Format Selection: LLMs were allowed to autonomously select the most appropriate format for each task, resulting in efficiency gains. This approach yielded a 3.3% to 5.7% improvement in reasoning tasks and reduced token usage by up to 72.7% in multi-agent communication while maintaining effectiveness.
  3. Methodology and Experiments:
    • The authors implemented a mechanism prompting LLMs to explore non-NL formats for task inputs. Through this, the capability of LLMs to autonomously decide the thought and communicative format was analyzed.
    • Performance improvements were quantified across both single-LLM reasoning tasks and multi-agent communication scenarios.
  4. Thought and Communication Transferability: The paper demonstrates that formats devised for reasoning and communication are transferable across different LLM architectures. The selected formats exhibited structural similarities to traditional agent communication languages, reflecting a spontaneous alignment with established communication protocols.
  5. Structural Formats in Communication: Interestingly, the structured formats selected by LLMs resemble traditional Agent Communication Languages (ACLs), such as KQML, highlighting the structured and efficient nature of these non-NL formats.
  6. Implications for ACLs: The resultant structured formats proposed by LLMs indicate a natural evolution toward efficient structured communication, underscoring the potential for developing new ACLs that leverage the strengths of LLM-autonomously selected formats.

The research underscores the feasibility and potential benefits of adopting non-NL formats for improving LLM functionalities in reasoning and inter-agent communication. This suggests a reevaluation of traditional NL reliance, highlighting the adaptability and efficiency of structured formats that could enrich agent communications and enhance computational efficiency. The code used in the experimentation is available online, facilitating further exploration and validation by the research community.

X Twitter Logo Streamline Icon: https://streamlinehq.com
Youtube Logo Streamline Icon: https://streamlinehq.com