Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Bridging the Gap: Deciphering Tabular Data Using Large Language Model (2308.11891v2)

Published 23 Aug 2023 in cs.CL and cs.AI

Abstract: In the realm of natural language processing, the understanding of tabular data has perpetually stood as a focal point of scholarly inquiry. The emergence of expansive LLMs, exemplified by the likes of ChatGPT, has ushered in a wave of endeavors wherein researchers aim to harness these models for tasks related to table-based question answering. Central to our investigative pursuits is the elucidation of methodologies that amplify the aptitude of such LLMs in discerning both the structural intricacies and inherent content of tables, ultimately facilitating their capacity to provide informed responses to pertinent queries. To this end, we have architected a distinctive module dedicated to the serialization of tables for seamless integration with expansive LLMs. Additionally, we've instituted a corrective mechanism within the model to rectify potential inaccuracies. Experimental results indicate that, although our proposed method trails the SOTA by approximately 11.7% in overall metrics, it surpasses the SOTA by about 1.2% in tests on specific datasets. This research marks the first application of LLMs to table-based question answering tasks, enhancing the model's comprehension of both table structures and content.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Hengyuan Zhang (34 papers)
  2. Peng Chang (13 papers)
  3. Zongcheng Ji (4 papers)
Youtube Logo Streamline Icon: https://streamlinehq.com