Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Symbol-LLM: Towards Foundational Symbol-centric Interface For Large Language Models (2311.09278v2)

Published 15 Nov 2023 in cs.CL and cs.AI

Abstract: Although LLMs demonstrate remarkable ability in processing and generating human-like text, they do have limitations when it comes to comprehending and expressing world knowledge that extends beyond the boundaries of natural language(e.g., chemical molecular formula). Injecting a collection of symbolic data directly into the training of LLMs can be problematic, as it disregards the synergies among different symbolic families and overlooks the need for a balanced mixture of natural and symbolic data. In this work, we tackle these challenges from both a data and framework perspective and introduce Symbol-LLM series models. First, we curated a data collection consisting of 34 tasks and incorporating approximately 20 distinct symbolic families, intending to capture the interrelations and foster synergies between symbols. Then, a two-stage tuning framework succeeds in injecting symbolic knowledge without loss of the generality ability. Extensive experiments on both symbol- and NL-centric tasks demonstrate the balanced and superior performances of Symbol-LLM series models. The project page is https://xufangzhi.github.io/symbol-LLM-page/.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Fangzhi Xu (22 papers)
  2. Zhiyong Wu (171 papers)
  3. Qiushi Sun (26 papers)
  4. Siyu Ren (24 papers)
  5. Fei Yuan (28 papers)
  6. Shuai Yuan (68 papers)
  7. Qika Lin (24 papers)
  8. Yu Qiao (563 papers)
  9. Jun Liu (606 papers)
Citations (21)
Github Logo Streamline Icon: https://streamlinehq.com