Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
38 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Making LLMs Work for Enterprise Data Tasks (2407.20256v1)

Published 22 Jul 2024 in cs.DB, cs.AI, and cs.LG

Abstract: LLMs know little about enterprise database tables in the private data ecosystem, which substantially differ from web text in structure and content. As LLMs' performance is tied to their training data, a crucial question is how useful they can be in improving enterprise database management and analysis tasks. To address this, we contribute experimental results on LLMs' performance for text-to-SQL and semantic column-type detection tasks on enterprise datasets. The performance of LLMs on enterprise data is significantly lower than on benchmark datasets commonly used. Informed by our findings and feedback from industry practitioners, we identify three fundamental challenges -- latency, cost, and quality -- and propose potential solutions to use LLMs in enterprise data workflows effectively.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Çağatay Demiralp (38 papers)
  2. Fabian Wenz (3 papers)
  3. Peter Baile Chen (9 papers)
  4. Moe Kayali (9 papers)
  5. Nesime Tatbul (20 papers)
  6. Michael Stonebraker (23 papers)