Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
97 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

A Unified Transferable Model for ML-Enhanced DBMS (2105.02418v3)

Published 6 May 2021 in cs.DB and cs.AI

Abstract: Recently, the database management system (DBMS) community has witnessed the power of ML solutions for DBMS tasks. Despite their promising performance, these existing solutions can hardly be considered satisfactory. First, these ML-based methods in DBMS are not effective enough because they are optimized on each specific task, and cannot explore or understand the intrinsic connections between tasks. Second, the training process has serious limitations that hinder their practicality, because they need to retrain the entire model from scratch for a new DB. Moreover, for each retraining, they require an excessive amount of training data, which is very expensive to acquire and unavailable for a new DB. We propose to explore the transferabilities of the ML methods both across tasks and across DBs to tackle these fundamental drawbacks. In this paper, we propose a unified model MTMLF that uses a multi-task training procedure to capture the transferable knowledge across tasks and a pre-train fine-tune procedure to distill the transferable meta knowledge across DBs. We believe this paradigm is more suitable for cloud DB service, and has the potential to revolutionize the way how ML is used in DBMS. Furthermore, to demonstrate the predicting power and viability of MTMLF, we provide a concrete and very promising case study on query optimization tasks. Last but not least, we discuss several concrete research opportunities along this line of work.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Ziniu Wu (20 papers)
  2. Pei Yu (45 papers)
  3. Peilun Yang (1 paper)
  4. Rong Zhu (34 papers)
  5. Yuxing Han (40 papers)
  6. Yaliang Li (117 papers)
  7. Defu Lian (142 papers)
  8. Kai Zeng (47 papers)
  9. Jingren Zhou (198 papers)
Citations (26)

Summary

We haven't generated a summary for this paper yet.