Papers
Topics
Authors
Recent
Search
2000 character limit reached

Can AI Agents Answer Your Data Questions? A Benchmark for Data Agents

Published 21 Mar 2026 in cs.DB | (2603.20576v1)

Abstract: Users across enterprises increasingly rely on AI agents to query their data through natural language. However, building reliable data agents remains difficult because real-world data is often fragmented across multiple heterogeneous database systems, with inconsistent references and information buried in unstructured text. Existing benchmarks only tackle individual pieces of this problem -- e.g., translating natural-language questions into SQL queries, answering questions over small tables provided in context -- but do not evaluate the full pipeline of integrating, transforming, and analyzing data across multiple database systems. To fill this gap, we present the Data Agent Benchmark (DAB), grounded in a formative study of enterprise data agent workloads across six industries. DAB comprises 54 queries across 12 datasets, 9 domains, and 4 database management systems. On DAB, the best frontier model (Gemini-3-Pro) achieves only 38% pass@1 accuracy. We benchmark five frontier LLMs, analyze their failure modes, and distill takeaways for future data agent development. Our benchmark and experiment code are published at github.com/ucbepic/DataAgentBench.

Summary

  • The paper introduces DAB as a novel benchmark that evaluates AI agents on complex multi-database integration and unstructured text extraction tasks.
  • The study systematically perturbs real datasets across multiple DBMSes to simulate enterprise challenges like schema mismatches and cost-accuracy trade-offs.
  • Evaluation of frontier LLM agents reveals key failure modes in planning and implementation, prompting the need for enhanced extraction tools and semantic layers.

Authoritative Summary of "Can AI Agents Answer Your Data Questions? A Benchmark for Data Agents" (2603.20576)

Motivation and Benchmark Design

This paper introduces the Data Agent Benchmark (DAB), targeting the evaluation of AI agents on end-to-end data analysis tasks representative of enterprise workloads. The authors identify a pronounced gap between existing evaluation methodologies, which are constrained to tasks such as text-to-SQL or Table-QA, and the real-world requirements for data agents. Enterprise data is distributed across heterogeneous database systems (e.g., SQL, NoSQL), with inconsistent references (e.g., ill-formatted join keys) and requires transformation of unstructured text. DAB is designed to encapsulate these complexities through (i) multi-database integration, (ii) ill-formatted join keys, (iii) unstructured text transformation, and (iv) domain knowledge requirements, as derived from a formative study across six industries. Figure 1

Figure 1

Figure 1: DAB architecture โ€“ agents operate in a ReAct loop, interacting with querying and Python tools, performing multi-database integration and unstructured text extraction.

Dataset Construction and Perturbation Methodology

DAB comprises 54 queries spanning 12 open-source datasets, 9 domains, and 4 DBMSes. The construction process involves systematic perturbations of clean datasets: columns that trivially answer queries are removed and re-embedded into free-text fields, and entity identifiers are reformatted to induce join mismatches. Realistic scenarios from the formative study are emulated by distributing data across PostgreSQL, SQLite, DuckDB, and MongoDB, necessitating dialect adaptation and reconciliation of schemas. Figure 2

Figure 2: The dataset creation pipeline for DAB, illustrating open data collection, transformations (key reformatting, free-text embedding), and heterogeneous DBMS distribution.

Evaluation of Frontier LLM Agents

Five frontier LLMsโ€”GPT-5.2, GPT-5-mini, Gemini-3-Pro, Gemini-2.5-Flash, Kimi-K2โ€”are benchmarked utilizing a ReAct-based agent loop, which iteratively reasons, calls tools (for querying, Python execution), and returns answers. Each agent is provided with logical database descriptions and hints for each dataset. 50 independent trials per query per agent are executed to estimate pass@kk; pass@1 denotes the probability of a correct answer in a single trial.

Accuracy and Cost Analysis

Gemini-3-Pro achieves the highest accuracy (38% pass@1) but at 20ร—\times the cost of GPT-5-mini, which achieves 30% pass@1 with the most favorable cost-accuracy tradeoff. No agent solves all queries in the "patents" dataset, and even with 50 attempts, best pass@50 does not exceed 69%. GPT-5-mini outperforms GPT-5.2 in accuracy despite being smaller, indicating that agent-specific strategy may outweigh model scale in this domain. Figure 3

Figure 3

Figure 3: Cost (log-scale USD) vs. pass@1โ€”GPT-5-mini optimizes cost-accuracy; Gemini-3-Pro achieves the highest accuracy at substantial cost. Pass@kk curve shows accuracy plateaus well below 100% even with k=50k=50.

Trajectory and Tool Use Diagnostics

Agents differ in strategy, with GPT-5-mini prioritizing SQL-based aggregation and minimizing Python calls, incurring low cost and high efficiency. Kimi-K2, on the other hand, issues broad queries and performs data processing in Python, resulting in higher computational and monetary overhead. Agents exploring schemas and data too little (e.g., Gemini-2.5-Flash) or too much (e.g., Kimi-K2) underperform relative to those maintaining moderate exploration rates (~20% of tool calls).

Failure Mode Taxonomy and Error Analysis

The authors define five primary failure modes: (FM1) fails before planning, (FM2) incorrect solution plan, (FM3) incorrect data selection, (FM4) incorrect implementation, (FM5) runtime error. Analysis across 1,147 trajectories finds that FM2 and FM4 account for ~85% of all failures, indicating that agents generally locate correct data but fail primarily in logical planning and implementation. Notably, every agent relies exclusively on regex-based extraction for free-text parsing and fails on tasks where regex is insufficient (e.g., varied date formats in patents). NLP-based or LLM-based extraction operators are not actively utilized, signifying a critical gap in agent tool repertoire. Figure 4

Figure 4: Trajectory breakdown by failure mode โ€“ most failures fall into FM2 (planning errors) and FM4 (implementation errors), runtime errors are rare.

Figure 5

Figure 5: Decomposition of planning and implementation failures โ€“ FM2 (incorrect plans) and FM4 (implementation errors) dominate; wrong data selection (FM3) is rare.

Case Study: PromptQL Production Agent

A case study with PromptQL, a commercial data agent platform, demonstrates that a semantic layer for metadata profiling and curated table relationships improves pass@1 by 7 percentage points over the ReAct baseline with the same backbone LLM. However, both architectures fail on queries requiring advanced extraction from unstructured text, reaffirming the bottleneck is in tool capability rather than data or schema exploration.

Comparative Analysis with Prior Benchmarks

DAB is contrasted with prior benchmarks (text-to-SQL, Table-QA, data science, semantic query processing, tool-use) and shown to uniquely encapsulate all four key properties (multi-database integration, ill-formatted keys, unstructured transformation, domain knowledge). Existing benchmarks are fundamentally limited by context (e.g., flat files, single database, structured data) and lack rigorous end-to-end workflow evaluation.

Practical and Theoretical Implications

DAB reveals unequivocally that frontier LLMs and agent frameworks are insufficient for robust enterprise data agent tasks, especially those involving cross-database reconciliation and semantic unstructured data processing. The dominant limitations are (i) poor logical planning, (ii) fragile implementation (esp. regex-reliance), and (iii) inadequate extraction tools. These findings motivate the development of agent frameworks supporting richer extraction primitives (NER, semantic parsing), semantic layers for schema abstraction, and improved planning modules. The explicit inclusion of domain knowledge and realistic data integration scenarios positions DAB as a critical instrument for future progress in trustworthy data agent systems.

Future Directions

Advanced data agents will require hybrid extraction modules combining symbolic and LLM-based text parsing, noise-robust planning, and adaptive exploration strategies. The integration of semantic query processing [patel2025semantic] [liu2025palimpzest] [jo2024thalamusdb], declarative extraction operators, and domain-adaptive prompting underscores the roadmap for practical deployments in heterogeneous enterprise environments. The low pass@1 observed even in optimal settings necessitates algorithmic innovations and rethinking agent toolchains.

Conclusion

The DAB benchmark constitutes a rigorous, production-grounded evaluation of data agent reliability and capability. The empirical findings demonstrate the substantial gap between current agentic tooling and realistic enterprise demands, especially on end-to-end heterogeneous workloads. The most salient error sources are agent planning and implementation failures, positioning DAB as foundational for future research in robust, production-oriented AI data agents.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

Explain it Like I'm 14

What this paper is about

This paper asks a simple question with a hard answer: Can todayโ€™s AI โ€œdata agentsโ€ really answer peopleโ€™s questions about company data just by using everyday language? To find out, the authors built a new test called the Data Agent Benchmark (DAB). It checks whether AI agents can do end-to-end data workโ€”finding, cleaning, combining, and analyzing real data spread across different systemsโ€”like a human analyst would.

What the researchers wanted to find out

The team focused on a few easy-to-understand questions:

  • Can AI agents handle data that lives in many different places and formats?
  • Can they match records when IDs or names donโ€™t line up (like โ€œabc_123โ€ in one table vs. โ€œid-123โ€ in another)?
  • Can they pull facts out of messy text (like finding a location or rating hidden inside a paragraph)?
  • Can they use domain knowledge (like finance or medical know-how) when the data alone isnโ€™t enough?
  • Overall, how accurate and reliable are these agents when asked realistic questions?

How they built and tested it

First, the team interviewed companies across six industries to learn what real user questions look like. Then they built DAB from open-source datasets (so everyone can use it) and carefully โ€œmessed them upโ€ to mimic real-world problems. Think of it like making a realistic obstacle course for data agents:

  • Multiple databases: Data is split across different systems (e.g., PostgreSQL, MongoDB, SQLite, DuckDB). An agent has to pull pieces from the right places and combine them.
  • Messy matching (ill-formatted join keys): The same thing might be labeled differently in each database, like having slightly different IDs or added spaces, so the agent has to recognize theyโ€™re the same and join them correctly.
  • Hidden facts in text (unstructured text): Useful details are buried in long text fields (like a review) instead of a neat column. The agent must extract and organize that info before using it.
  • Domain knowledge: Some questions need expert rules (e.g., how to compute a financial metric correctly or what a medical term means).

They created 54 natural-language questions across 12 datasets and 9 topic areas (like news, books, CRM, GitHub, Yelp, stocks, patents, and cancer research). Each question has a clear, correct answer that can be checked by code.

To test AI agents, they used a common approach called a ReAct loop. Picture it like this:

  1. the agent thinks about the next step,
  2. runs a tool (like making a database query or a small Python script),
  3. looks at the result,
  4. decides what to do next, and repeats until it has an answer.

They evaluated five strong LLMs acting as agents, each allowed to:

  • list what databases and tables exist,
  • run read-only queries,
  • execute small Python snippets,
  • and return a final answer.

They tried each question 50 times to account for randomness and measured accuracy using โ€œpass@1โ€ (getting it right on the first try) and โ€œpass@50โ€ (getting it right at least once across 50 tries).

What they discovered

Results were surprisingly low. Even the best model (Gemini-3-Pro) got only 38% of questions right on the first try (pass@1). Even after 50 tries, it stayed under 69%. One dataset (patents) wasnโ€™t solved correctly by any agent in any trial.

Here are the key takeaways, explained simply:

  • End-to-end is hard: Agents struggled with the full pipelineโ€”finding the right tables, cleaning keys, extracting from text, and applying domain rulesโ€”not just writing a single SQL query.
  • A โ€œjust rightโ€ amount of exploring helps: The best-performing agents used about 20% of their actions to explore the data (like checking schemas or peeking at sample rows). Too little exploring misses important details; too much wastes time.
  • Most errors were planning or implementation mistakes: About 85% of wrong answers came from flawed plans or buggy steps (e.g., not cleaning IDs correctly or misusing a query), not from picking the wrong database.
  • Text extraction was basic: Every agent relied on simple patterns, like regular expressions (regex), to pull data out of text. None used more advanced NLP or LLM-based extraction inside the workflow.
  • A production system helped a bit, but not enough: A specialized agent from an industry partner (PromptQL) improved first-try accuracy by 7 percentage points over the baseline setup but still failed completely on questions that required extracting data from unstructured text.

Why this matters

Many companies hope to let employees ask questions about their data in plain language. This paper shows weโ€™re not there yet: real-world data is split across systems, has mismatched labels, and hides key facts inside text. Current AI agents often stumble on these specifics.

But the benchmark itself (DAB) is a big step forward. It gives researchers and companies a realistic, repeatable way to measure progress on end-to-end data tasks, not just pieces of the problem. The findings suggest clear directions for improvement:

  • Better tools for extracting structured info from messy text (beyond simple regex).
  • Smarter planning and checkingโ€”so agents can design and follow correct multi-step plans.
  • Helpful โ€œsemantic layersโ€ or shared definitions that reduce planning burden (e.g., standard ways to compute domain-specific metrics).
  • Balanced exploration strategies to discover whatโ€™s in the data without getting lost.

In short, DAB highlights whatโ€™s hard about real data work and gives the community a common challenge to push AI data agents toward being truly useful in everyday business settings.

Knowledge Gaps

Knowledge gaps, limitations, and open questions

The following list summarizes what remains missing, uncertain, or unexplored, with concrete directions for future work:

  • Benchmark scope excludes API integration and open-ended analytical reasoning. Develop extensions that: (a) evaluate API usage via snapshot/mock endpoints with deterministic outputs; and (b) assess open-ended tasks using rubric-based or programmatic judges (e.g., metric sets, constraint checks) that preserve reproducibility.
  • Reliance on dataset-level hints. Measure agent performance with varied levels of assistance (no hints, partial, noisy, conflicting), and provide DAB variants to isolate robustness to missing or imperfect domain context.
  • Limited DBMS coverage. Add enterprise systems (Snowflake, BigQuery, SQL Server, MySQL) and assess dialect-specific challenges (e.g., window functions, UDFs, semi-structured types, stored procedures, query optimizers).
  • Stylized data perturbations may not fully capture real-world messiness. Validate representativeness by: (a) parameterizing corruption severity; (b) incorporating more realistic fuzzy entity resolution (synonyms, abbreviations, transpositions, multi-field matching); and (c) conducting external audits against production logs or red-teamed datasets.
  • Read-only tasks only. Introduce controlled write scenarios (e.g., cleaning, normalization, materialized views) with guardrails and rollback, to evaluate agentsโ€™ ability to propose and carry out corrective actions safely.
  • Evaluation favors recall over precision. Design automatic graders with structured output schemas, canonicalization/normalization, numeric tolerance, and explicit checks for extraneous itemsโ€”without resorting to LLM judgesโ€”to better penalize over-inclusive answers.
  • Benchmark scale and coverage. Expand the number and diversity of queries, include multi-hop analytical chains, and provide per-property difficulty tiers and train/dev/test splits to support method development and reliable comparison.
  • Architecture diversity is not explored. Systematically compare agent frameworks beyond ReAct (e.g., planning/decomposition, workflow orchestration, self-correction, memory-enabled agents, multi-agent collaboration) under matched toolsets and prompts.
  • Tooling is minimal (list_db, query_db, execute_python). Evaluate the impact of richer primitives: row-wise semantic operators (LLM apply), built-in extractors (regex builder, pattern libraries), schema search/summarization, data profilers, and join-assist utilities.
  • Model configuration and sensitivity. Study effects of temperature, reasoning effort, tool-call batching/parallelism, context window size, and prompt variants; include additional frontier/open models (e.g., Claude) and report standardized seeds and variance.
  • Explorationโ€“exploitation trade-offs are observed but not optimized. Develop adaptive controllers to tune exploration budgets (e.g., schema/data inspection rate) and quantify how exploration allocation affects accuracy, latency, and cost.
  • Context management is intentionally minimal. Investigate structured memory, automatic summarization of tool results, chunked retrieval, and caching strategies; quantify the effect of truncation policies on downstream planning and correctness.
  • Security and safety are not evaluated. Test for prompt injection via database content, unsafe code execution in Python, data exfiltration, privilege misuse, and sandbox escapes; define and benchmark mitigations (e.g., taint tracking, policy guards, network isolation).
  • Single-turn tasks only. Add multi-turn interaction tasks that require clarifications, reformulations, and memory of prior context; measure agentsโ€™ ability to ask for missing information and maintain state across turns.
  • Scalability limits untested. Introduce tasks over large tables/corpora (tens of millions of rows), streaming or time-partitioned data, and complex join graphs; report latency/cost scaling and success under strict time budgets.
  • Ground-truth alignment after transformations. Provide automated consistency checks that ensure corrupted/perturbed datasets still deterministically yield the same answers; release regeneration scripts and versioned artifacts to support auditing.
  • No detailed accuracy breakdown by benchmark property. Report per-property and cross-property performance (multi-db integration, ill-formatted keys, text extraction, domain knowledge) to identify which challenges drive failure.
  • Fuzzy ER underrepresented. Include tasks requiring approximate string matching, multi-attribute linkage, probabilistic resolution, and deduplication across sources; benchmark ER algorithms as callable tools.
  • Domain knowledge is supplied via hints; retrieval is not tested. Add settings where agents must fetch background knowledge from documentation or web snapshots and integrate it with database content under reproducible constraints.
  • The patents dataset is entirely unsolved. Provide a deeper error analysis and targeted sub-tasks (e.g., simpler CPC hierarchy navigation, intermediate scaffolding) to isolate failure causes and incremental improvements.
  • Cost reporting focuses on USD totals. Include token-level accounting, latency breakdown per tool, and costโ€“accuracy trade-off curves; explore budgeted pass@k optimization and anytime stopping policies.
  • pass@k independence assumption is unverified. Report trial-level variance, correlations across attempts, and confidence intervals; consider bootstrapping or controlled randomization to validate metric assumptions.
  • Reproducibility across model updates is not addressed. Establish model/version pinning, release prompt and tool specs with hashes, and track performance drift over time.
  • Schema evolution and data volatility are absent. Add tasks requiring schema discovery across versions, handling missing/renamed columns, and time-aware reasoning (e.g., โ€œassume today is โ€ฆโ€).
  • MongoDB coverage is limited. Include aggregation pipelines, nested arrays/documents, and cross-collection joins; evaluate translation challenges between SQL and document-oriented paradigms.
  • Bias and calibration in text-derived labels are not analyzed. For classification/extraction tasks, measure calibration, error types, and fairness across categories; provide ground-truth distributions to detect systematic biases.

Practical Applications

Immediate Applications

Below are practical, deployable-now applications that leverage the benchmark, methods, and findings reported in the paper.

  • Procurement-grade evaluation of data agents for enterprises
    • Sectors: software/SaaS, finance, healthcare, e-commerce, analytics vendors
    • What: Use DAB to pressure-test vendor claims on end-to-end data agents (multi-DB integration, ill-formatted joins, text extraction, domain knowledge) with pass@k and cost metrics. Establish acceptance thresholds and compare costโ€“accuracy tradeoffs (e.g., GPT-5-mini vs. Gemini-3-Pro).
    • Tools/workflows: Internal โ€œAgent QA Suiteโ€ built around DAB; CI pipelines running nightly pass@1/pass@k; dashboards showing accuracy by dataset and failure taxonomy.
    • Dependencies/assumptions: Access to model/tool-calling APIs; sandboxed DB instances; budget for evaluation runs.
  • Pre-deployment regression testing and SLO setting for in-house data agents
    • Sectors: any enterprise with analytics agents; data platform teams
    • What: Adopt DAB as a regression suite to catch planning and implementation regressions before production. Use findings on exploration behavior (โ‰ˆ20% tool calls on schema/data exploration) to tune agent prompts and guardrails.
    • Tools/workflows: GitHub Actions + DAB; automatic diffing of pass@k, latency, cost per release; heuristic limits on exploration vs. execution steps.
    • Dependencies/assumptions: Stable staging data; telemetry to measure exploration ratios.
  • Cost-aware model selection and routing
    • Sectors: data platforms, BI tools, ML infra
    • What: Route low-stakes or simple queries to lower-cost models (e.g., GPT-5-mini) and high-complexity ones to higher-accuracy models, guided by DABโ€™s costโ€“accuracy curves.
    • Tools/workflows: Query complexity classifier; model-routing policy based on expected pass@1 and cost.
    • Dependencies/assumptions: Historical logs to train a router; monitoring to detect drift.
  • Ready-to-use ReAct agent scaffolding with secure tool execution
    • Sectors: software, platforms
    • What: Reuse the paperโ€™s ReAct loop, tool set (list_db, query_db, execute_python, return_answer), and safety practices (read-only query enforcement, timeouts, context truncation, result file handles).
    • Tools/workflows: Agent SDK templates; Dockerized Python sandbox with Pandas/PyArrow; SQL/Mongo read-only checkers.
    • Dependencies/assumptions: Secure execution environment; DB credentials management.
  • Operational guardrails and observability based on failure taxonomy
    • Sectors: enterprise IT, analytics, regulated industries
    • What: Instrument agents to log plan/implementation errors vs. source selection errors; alert when unstructured-text extraction is attempted via brittle regex only.
    • Tools/workflows: Telemetry tags for error classes; dashboards with heatmaps by failure mode; automatic fallback to human review on risky modes.
    • Dependencies/assumptions: Consistent error labeling; analyst-on-call workflows.
  • Off-the-shelf โ€œJoin-Key Reconcilerโ€ microservice for cross-DB joins
    • Sectors: data engineering, CRMs, e-commerce
    • What: Package deterministic join-key fixes (prefix stripping, whitespace cleanup, stable mappings) discovered in DAB into a callable service to standardize identifiers before queries.
    • Tools/workflows: Pre-join transform pipeline; library of deterministic transforms; config for dataset-specific rules.
    • Dependencies/assumptions: Access to mappings; stable patterns in identifier corruption.
  • Lightweight text-extraction primitives for agents (regex + patterns)
    • Sectors: software, support/CRM analytics, finance
    • What: Provide agents with a curated library of safe, tested extraction patterns for common fields (timestamps, ratings, locations) to reduce reliance on ad-hoc regex in the agentโ€™s prompt.
    • Tools/workflows: โ€œExtraction Toolkitโ€ (pattern catalog, validators); unit tests per pattern; integration with execute_python.
    • Dependencies/assumptions: Coverage for frequent patterns; versioning and regression tests.
  • Semantic layer documentation surfaced to the agent
    • Sectors: BI, finance, healthcare
    • What: Expose business definitions, metric formulas, and known mappings (as in DAB hints) to reduce planning burden and improve correctness in domain-knowledge tasks.
    • Tools/workflows: Metrics catalog (e.g., YAML or dbt docs) injected as context; agent-side retrieval of definitions prior to planning.
    • Dependencies/assumptions: Maintained semantic layer; governance for definitions.
  • Academic use in courses and research labs
    • Sectors: academia (CS, data systems, HCI)
    • What: Use DAB to teach end-to-end data-centric agents, multi-DB integration, and evaluation. Run ablations on exploration budgets and tool design.
    • Tools/workflows: Course assignments; reproducible notebooks; leaderboards for student agents.
    • Dependencies/assumptions: Computing credits; containerized DBMS setups.
  • Policy and risk assessment checklists for AI-driven analytics
    • Sectors: public sector, healthcare, finance
    • What: Create procurement and risk checklists requiring DAB-style end-to-end accuracy evidence, read-only enforcement, audit logs of tool calls, and cost reporting.
    • Tools/workflows: Standardized reporting templates (pass@k, dataset coverage, failure modes); internal approvals before deployment.
    • Dependencies/assumptions: Organizational mandate for evaluation; data privacy constraints.
  • Human-in-the-loop workflows for high-stakes queries
    • Sectors: healthcare, finance, operations
    • What: Given low pass@1 (โ‰ˆ38% best), route multi-DB and unstructured-text tasks to analyst review at key steps (plan approval, join-key mapping, final answer).
    • Tools/workflows: UI for plan preview; stepwise confirmations; diff views of intermediate query results.
    • Dependencies/assumptions: Trained reviewers; SLAs acknowledging added latency.
  • Domain-specific sandboxes for safe experimentation
    • Sectors: healthcare, finance, IP/patents
    • What: Mirror DABโ€™s dataset distribution to create internal โ€œmessy mirrorsโ€ of production schemas (with synthetic perturbations) for safe agent testing.
    • Tools/workflows: Data virtualization of production metadata; seeded perturbations (key reformatting, text embedding); deterministic ground truths.
    • Dependencies/assumptions: Synthetic or de-identified data; reproducibility infrastructure.

Long-Term Applications

These applications require further research, scaling, or productization beyond what the current results support.

  • Robust, production-grade data agents for multi-DB analytics
    • Sectors: BI, SaaS/CRM, finance, healthcare, e-commerce
    • What: Agents that reliably integrate heterogeneous backends, reconcile identifiers, transform unstructured text, and apply domain definitions at scale.
    • Potential products: โ€œEnterprise Data Copilotโ€ with certified pass@k thresholds per domain.
    • Dependencies/assumptions: Significant improvements in planning and implementation reliability; stronger text-extraction methods than regex; organizational adoption of semantic layers.
  • Database-native semantic operators for text (LLM UDFs)
    • Sectors: DBMS vendors, enterprise data platforms
    • What: Safe, cost-controlled row-wise LLM operators (classify/extract/cluster/summarize/search) exposed in SQL and integrated with query planners.
    • Potential products: SQL extensions (e.g., EXTRACT_JSON(field USING MODEL โ€ฆ)) with budgets and caching; vectorization across rows.
    • Dependencies/assumptions: Execution cost controls, caching, and data privacy controls; accuracy monitoring.
  • Automated semantic layer discovery and maintenance
    • Sectors: analytics, governance
    • What: Learn or mine business metrics, entity mappings, and synonyms from code, docs, and query logs to autopopulate the โ€œhintsโ€ agents need.
    • Potential products: โ€œSemantic Layer Builderโ€ that suggests canonical joins, metric formulas, and abbreviations.
    • Dependencies/assumptions: Access to documentation and logs; human validation loops; versioning and lineage.
  • Learned entity resolution and fuzzy joining across systems
    • Sectors: CRM, supply chain, healthcare
    • What: Replace deterministic transforms with learned matching models that resolve entities across noisy systems (names, addresses, abbreviations).
    • Potential products: โ€œCross-DB Resolverโ€ service with active learning and confidence thresholds integrated into agent workflows.
    • Dependencies/assumptions: Labeled pairs or weak supervision; privacy-compliant training; performance guarantees and fallbacks.
  • Planning-first agent architectures with verification
    • Sectors: software platforms, research
    • What: Agents that synthesize and verify executable plans before running tools, with unit tests on sub-steps and self-correction when checks fail.
    • Potential products: Plannerโ€“Executor frameworks; program-synthesis with static analysis; query plan validation against small samples.
    • Dependencies/assumptions: Advances in program synthesis and verifiable reasoning; tight integration with data sampling and test oracles.
  • Cross-DB interoperability layers for agents
    • Sectors: data engineering, vendors
    • What: Standardized intermediate representation (IR) for multi-DB queries and tool calls, enabling optimization and provenance across heterogeneous systems.
    • Potential products: Agent Query IR + adapters for SQL dialects and MongoDB; cross-engine cost models.
    • Dependencies/assumptions: Community standards; performance benchmarks; security and governance alignment.
  • Evaluation standards and certification for AI data agents
    • Sectors: regulators, standards bodies, large enterprises
    • What: Formalize DAB-like end-to-end evaluations as certifications for data agents (accuracy, cost, auditability), tailored to risk tiers.
    • Potential products: โ€œData Agent Safety Markโ€ with test batteries per sector (e.g., healthcare addenda).
    • Dependencies/assumptions: Multi-stakeholder consensus; periodic test refresh to prevent overfitting.
  • Personal knowledge agents that integrate across consumer data silos
    • Sectors: consumer productivity, privacy tech
    • What: Agents that reason over emails, documents, calendars, and notes to answer questions requiring multi-source joins and text extraction.
    • Potential products: โ€œPersonal Data Analystโ€ with local-first execution and privacy-preserving sandboxes.
    • Dependencies/assumptions: Strong privacy/security; reliable unstructured-text extraction and entity reconciliation; app API access.
  • Sector-specific copilots with domain knowledge libraries
    • Sectors: healthcare (EHR + notes), finance (market + internal risk data), energy (sensor + work orders)
    • What: Pre-packaged agents with curated domain metrics, mappings, and extraction patterns; validated against sector-specific DAB variants.
    • Potential products: โ€œCRM Success Copilot,โ€ โ€œClinical Trial Data Copilot,โ€ โ€œPlant Ops Copilot.โ€
    • Dependencies/assumptions: Curated, up-to-date domain ontologies; compliance approvals; ongoing evaluation and monitoring.
  • Humanโ€“AI collaboration loops for dataset and policy evolution
    • Sectors: enterprise IT, governance
    • What: Agents propose join rules/extraction patterns; humans approve and promote to the semantic layer; policies evolve with audit trails.
    • Potential products: โ€œAgent Suggestion Inboxโ€ with diff-based reviews; auto-rollback on error spikes.
    • Dependencies/assumptions: UX for reviewers; metrics to detect regressions; cultural adoption.
  • Expanded research benchmarks (APIs, streaming, open-ended tasks)
    • Sectors: research, vendors
    • What: Extend DAB to include API integrations, streaming sources, and exploratory analytics with rubric-based judgments.
    • Potential products: DAB-API and DAB-Explore tracks; public leaderboards and workshops.
    • Dependencies/assumptions: Stable evaluation protocols for non-deterministic tasks; LLM-judge reliability or human review budgets.
  • Privacy- and compliance-aware agent runtimes
    • Sectors: healthcare, finance, public sector
    • What: Policy engines that enforce read-only modes, data minimization, per-source access policies, and comprehensive tool-call audit logs while maintaining agent performance.
    • Potential products: โ€œCompliance Agent Runtimeโ€ integrating with IAM, DLP, and SIEM.
    • Dependencies/assumptions: Integration with enterprise identity and data catalogs; legal sign-off and audits.

Notes on feasibility across items:

  • Current accuracy limits (best pass@1 โ‰ˆ 38%, pass@50 โ‰ค 69%) necessitate human oversight for high-stakes use.
  • Many long-term applications depend on advances in text extraction beyond regex, improved planning/verification, and robust semantic layers.
  • Secure, sandboxed execution and governance are prerequisites for both immediate and long-term deployments.

Glossary

  • adjusted closing prices: Price series corrected for corporate actions (splits, dividends) used to compute comparable returns and volatility. "must be computed from adjusted closing prices to account for splits and dividends."
  • agent trajectory: The full sequence of an agentโ€™s model calls, tool invocations, and observed results during a trial. "All agent trajectories are publicly available."
  • API integration: Combining data from external application programming interfaces with database sources within a single workflow. "API integration, where relevant data resides not in databases but in external APIs (e.g., email clients, web search endpoints, third-party data providers) that must be queried alongside database sources."
  • approximate string matching: Techniques that match similar but not identical strings (e.g., to align noisy identifiers). "using approximate string matching or contextual reasoning"
  • BRCA: A cancer type (Breast Invasive Carcinoma) designation used in genomics datasets and analyses. "Among alive BRCA patients, which top 3 histological types show the highest % of CDH1 gene mutations?"
  • CDH1 gene mutations: Alterations in the CDH1 gene (E-cadherin) with clinical and biological significance in certain cancers. "CDH1 gene mutations"
  • Cooperative Patent Classification (CPC): A hierarchical system for categorizing patents into technology areas. "Identify CPC areas with the highest EMA of patent filings (smoothing 0.2); return level-5 codes whose best year is 2022."
  • CRM (Customer Relationship Management): Systems and processes for managing interactions with customers, leads, and sales operations. "customer relationship management (CRM) tool"
  • data-dependent transformations: Text/data processing where rules vary per record and cannot be captured by a single fixed pattern. "Data-dependent transformations require the agent to examine individual rows"
  • data-independent transformations: Text/data processing solvable by a fixed program or pattern that applies uniformly to all rows. "Data-independent transformations can be resolved by fixed-size programs regardless of data cardinality."
  • data silos: Isolated data stores that hinder integration and cross-system analysis. "store data in disparate silos"
  • database management system (DBMS): Software for creating, querying, and managing databases. "4 database management systems (DBMSes)."
  • Docker environment: A containerized runtime that packages code and dependencies for consistent execution. "in a Docker environment with Python 3.12"
  • entity resolution: The process of determining when records from different sources refer to the same real-world entity. "where entity resolution is required to match records across sources"
  • exact-match joins: Joins where identifiers align perfectly across sources without transformation. "exact-match joins, where identifiers match one-to-one across sources"
  • Exponential Moving Average (EMA): A time-series smoothing technique that weights recent observations more heavily. "Identify CPC areas with the highest EMA of patent filings (smoothing 0.2)"
  • failure taxonomy: A structured categorization of error types or failure modes observed in system behavior. "We develop a failure taxonomy"
  • frontier LLMs: State-of-the-art LLMs at the leading edge of capability. "We benchmark five frontier LLMs"
  • fuzzy joins: Cross-source joins that rely on approximate matching or contextual cues due to inconsistent identifiers. "fuzzy joins, where entity resolution is required to match records across sources"
  • ground-truth answer: The correct, authoritative answer used to evaluate system outputs. "Each query consists of a natural-language question, a ground-truth answer, and a validation script."
  • histological types: Categories of tissue characteristics used to classify tumors in pathology. "which top 3 histological types show the highest % of CDH1 gene mutations?"
  • ill-formatted join keys: Identifiers for the same entity that differ across sources (e.g., prefixes, whitespace), requiring normalization before joining. "ill-formatted join keys"
  • intraday volatility: A measure of how much an assetโ€™s price fluctuates within the trading day. "Which stock index in Asia has the highest average intraday volatility since 2020?"
  • LLM-based judge: A LLM used as an evaluator to assess correctness or quality of outputs. "would require either manual inspection of every trajectory or an LLM-based judge"
  • open-ended analytical reasoning: Exploratory analysis where the approach and metrics are not pre-specified and multiple answers may be acceptable. "Open-ended analytical reasoning."
  • pass@1: The probability that a single attempt produces a correct answer (special case of pass@k with k=1). "achieves only 38% pass@1 accuracy"
  • pass@k: The probability that at least one of k independent attempts succeeds; a reliability metric for stochastic systems. "measure accuracy using pass@kk"
  • petabyte-scale: Data sizes on the order of 1015 bytes, indicating extremely large volumes. "petabyte-scale data volumes."
  • programmatic-transform joins: Cross-source joins where identifiers differ but can be reconciled via deterministic transformations. "programmatic-transform joins, where identifiers refer to the same entity but differ in format and can be reconciled via deterministic rules"
  • ReAct pattern: An agent design that alternates between reasoning and acting via tool calls in iterative loops. "using the ReAct pattern"
  • regular expressions: Pattern-matching syntax for extracting or transforming text based on specified rules. "Every agent uses regular expressions"
  • semantic layer: An intermediate abstraction that encodes business logic and concepts to simplify queries and planning. "semantic layers can reduce the planning burden on the agent."
  • semantic mapping: Mapping between semantically equivalent but differently represented concepts or identifiers. "require a semantic mapping between full stock exchange names and abbreviated index symbols"
  • semantic operators: LLM-powered row-level text operations such as classification, extraction, clustering, summarization, and semantic search. "semantic operatorsโ€”i.e., LLM-powered transformations applied to individual rows of a table"
  • SQL dialects: Variants of SQL syntax and features across different database systems. "SQL dialects and MongoDB's query language"
  • stratified: Averaging or analysis performed within defined groups (strata) before aggregating, to prevent disproportionate weighting. "All reported averages are stratified"
  • Table-QA: Table question answering; tasks where answers are derived from tables, often provided directly in context. "table question-answering (i.e., Table-QA) benchmarks"
  • text-to-SQL: Translating natural-language questions into executable SQL queries. "text-to-SQL benchmarks"
  • thematic analysis: A qualitative research method for identifying patterns (themes) in data. "thematic analysis, a widely used qualitative method in HCI research"
  • tool calling: LLM capability to invoke external functions/tools by emitting structured function calls from prompts. "Modern LLMs support tool calling"
  • unstructured text transformation: Converting free-form text into structured data (e.g., extracting fields) for downstream querying and analysis. "unstructured text transformation: answers may be embedded in text fields that the agent must parse into structured values"

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 415 likes about this paper.