Effective Distillation of Table-based Reasoning Ability from LLMs (2309.13182v2)
Abstract: LLMs have demonstrated remarkable performance across a wide range of natural language processing tasks. However, their enormous parameter size and extremely high requirements for compute power pose challenges for their practical deployment. Recent research has revealed that specific capabilities of LLMs, such as numerical reasoning, can be transferred to smaller models through distillation. Some studies explore the potential of leveraging LLMs to perform table-based reasoning. However, there has been no prior work focusing on table reasoning skills in smaller models specifically tailored for scientific table-to-text generation tasks. In this paper, we propose a novel table-based reasoning distillation approach, with the aim of distilling LLMs into tailored smaller models. Our experimental results have shown that a 220 million parameter model (Flan-T5-base) fine-tuned using distilled data, not only achieves a significant improvement compared to traditionally fine-tuned baselines, but also surpasses specific LLMs on a scientific table-to-text generation dataset. Our code is available at https://github.com/Bernard-Yang/DistillTableCoT.
- Feverous: Fact extraction and verification over unstructured and structured information. ArXiv, abs/2106.05707.
- Satanjeev Banerjee and Alon Lavie. 2005. Meteor: An automatic metric for mt evaluation with improved correlation with human judgments. In IEEvaluation@ACL.
- Language models are few-shot learners. ArXiv, abs/2005.14165.
- Star: Sql guided pre-training for context-dependent text-to-sql parsing. ArXiv, abs/2210.11888.
- Wenhu Chen. 2023. Large Language Models are few(1)-shot Table Reasoners. In Findings of the Association for Computational Linguistics: EACL 2023, pages 1120–1130, Dubrovnik, Croatia. Association for Computational Linguistics.
- Logical Natural Language Generation from Open-Domain Tables. ArXiv:2004.10404 [cs].
- TABFACT: A LARGE-SCALE DATASET FOR TABLE- BASED FACT VERIFICATION.
- Logic2Text: High-Fidelity Natural Language Generation from Logical Forms. arXiv:2004.14579 [cs]. ArXiv: 2004.14579.
- Binding Language Models in Symbolic Languages. ArXiv:2210.02875 [cs].
- Explanatory and actionable debugging for machine learning: A tableqa demonstration. Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval.
- Palm: Scaling language modeling with pathways. ArXiv, abs/2204.02311.
- Scaling instruction-finetuned language models. ArXiv, abs/2210.11416.
- Inducing and Using Alignments for Transition-based AMR Parsing. ArXiv:2205.01464 [cs].
- Specializing Smaller Language Models towards Multi-Step Reasoning. ArXiv:2301.12726 [cs].
- Carlos Gemmell and Jeffrey Stephen Dalton. 2023. Generate, transform, answer: Question specific tool synthesis for tabular data. ArXiv, abs/2303.10138.
- Infotabs: Inference on tables as semi-structured data. In Annual Meeting of the Association for Computational Linguistics.
- TAPAS: Weakly Supervised Table Parsing via Pre-training. arXiv:2004.02349 [cs]. ArXiv: 2004.02349.
- Distilling the knowledge in a neural network. ArXiv, abs/1503.02531.
- Large Language Models Are Reasoning Teachers. ArXiv:2212.10071 [cs].
- Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes. ArXiv, abs/2305.02301.
- Omnitab: Pretraining with natural and synthetic data for few-shot table-based question answering. In North American Chapter of the Association for Computational Linguistics.
- Large language models are zero-shot reasoners. ArXiv, abs/2205.11916.
- BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, Online. Association for Computational Linguistics.
- PLOG: Table-to-Logic Pretraining for Logical Table-to-Text Generation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 5531–5546, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- TAPEX: TABLE PRE-TRAINING VIA LEARNING A NEURAL SQL EXECUTOR.
- Towards Faithfulness in Open Domain Table-to-text Generation from an Entity-centric View. arXiv:2102.08585 [cs]. ArXiv: 2102.08585.
- SCITAB: A Challenging Benchmark for Compositional Reasoning and Claim Verification on Scientific Tables. ArXiv:2305.13186 [cs].
- Self-refine: Iterative refinement with self-feedback. arXiv preprint arXiv:2303.17651.
- Teaching Small Language Models to Reason. ArXiv:2212.08410 [cs].
- SciGen: a Dataset for Reasoning-Aware Text Generation from Scientific Tables.
- FeTaQA: Free-form Table Question Answering. Transactions of the Association for Computational Linguistics, 10:35–49.
- Enhancing few-shot text-to-sql capabilities of large language models: A study on prompt design strategies. arXiv preprint arXiv:2305.12586.
- Panupong Pasupat and Percy Liang. 2015. Compositional semantic parsing on semi-structured tables. In Annual Meeting of the Association for Computational Linguistics.
- Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. ArXiv.
- Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. ArXiv, abs/1910.01108.
- Bleurt: Learning robust metrics for text generation. In ACL.
- Llama: Open and efficient foundation language models. ArXiv, abs/2302.13971.
- Self-Consistency Improves Chain of Thought Reasoning in Language Models. ArXiv:2203.11171 [cs].
- Self-consistency improves chain of thought reasoning in language models. ArXiv, abs/2203.11171.
- Chain of thought prompting elicits reasoning in large language models. ArXiv, abs/2201.11903.
- Large Language Models are Versatile Decomposers: Decompose Evidence and Questions for Table-based Reasoning. ArXiv:2301.13808 [cs].
- Spider: A large-scale human-labeled dataset for complex and cross-domain semantic parsing and text-to-sql task. ArXiv, abs/1809.08887.
- Glm-130b: An open bilingual pre-trained model. ArXiv, abs/2210.02414.
- Bertscore: Evaluating text generation with bert. ArXiv, abs/1904.09675.
- Reastap: Injecting table reasoning skills during pre-training via synthetic reasoning examples. In Conference on Empirical Methods in Natural Language Processing.
- Large Language Models are Effective Table-to-Text Generators, Evaluators, and Feedback Providers. ArXiv:2305.14987 [cs].
- Seq2sql: Generating structured queries from natural language using reinforcement learning. ArXiv, abs/1709.00103.
- LEAST-TO-MOST PROMPTING ENABLES COMPLEX REASONING IN LARGE LANGUAGE MODELS.
- PaD: Program-aided Distillation Specializes Large Models in Reasoning. ArXiv:2305.13888 [cs].
- Bohao Yang (16 papers)
- Chen Tang (94 papers)
- Kun Zhao (97 papers)
- Chenghao Xiao (21 papers)
- Chenghua Lin (127 papers)