Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

TUTA: Tree-based Transformers for Generally Structured Table Pre-training (2010.12537v4)

Published 21 Oct 2020 in cs.IR, cs.AI, and cs.DB

Abstract: Tables are widely used with various structures to organize and present data. Recent attempts on table understanding mainly focus on relational tables, yet overlook to other common table structures. In this paper, we propose TUTA, a unified pre-training architecture for understanding generally structured tables. Noticing that understanding a table requires spatial, hierarchical, and semantic information, we enhance transformers with three novel structure-aware mechanisms. First, we devise a unified tree-based structure, called a bi-dimensional coordinate tree, to describe both the spatial and hierarchical information of generally structured tables. Upon this, we propose tree-based attention and position embedding to better capture the spatial and hierarchical information. Moreover, we devise three progressive pre-training objectives to enable representations at the token, cell, and table levels. We pre-train TUTA on a wide range of unlabeled web and spreadsheet tables and fine-tune it on two critical tasks in the field of table structure understanding: cell type classification and table type classification. Experiments show that TUTA is highly effective, achieving state-of-the-art on five widely-studied datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (7)
  1. Zhiruo Wang (18 papers)
  2. Haoyu Dong (55 papers)
  3. Ran Jia (7 papers)
  4. Jia Li (380 papers)
  5. Zhiyi Fu (10 papers)
  6. Shi Han (74 papers)
  7. Dongmei Zhang (193 papers)
Citations (120)

Summary

We haven't generated a summary for this paper yet.