Xiang Deng, Huan Sun, Alyssa Lees, You Wu, Cong Yu
Relational tables on the Web store a vast amount of knowledge. Owing to the wealth of such tables, there has been tremendous progress on a variety of tasks in the area of table understanding. However, existing work generally relies on heavily-engineered task-specific features and model architectures. In this paper, we present TURL, a novel framework that introduces the pre-training/fine-tuning paradigm to relational Web tables. During pre-training, our framework learns deep contextualized representations on relational tables in an unsupervised manner. Its universal model design with pre-trained representations can be applied to a wide range of tasks with minimal task-specific fine-tuning. Specifically, we propose a structure-aware Transformer encoder to model the row-column structure of relational tables, and present a new Masked Entity Recovery (MER) objective for pre-training to capture the semantics and knowledge in large-scale unlabeled data. We systematically evaluate TURL with a benchmark consisting of 6 different tasks for table understanding (e.g., relation extraction, cell filling). We show that TURL generalizes well to all tasks and substantially outperforms existing methods in almost all instances.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Data Integration | T2Dv2 | Accuracy (%) | 96.2 | TURL |
| Data Integration | WikiTables-TURL-CTA | F1 (%) | 94.75 | TURL |
| Data Integration | WikipediaGS-CTA | Accuracy (%) | 74.6 | TURL |
| Data Integration | WikiTables-TURL-CEA | F1 (%) | 68 | TURL |
| Data Integration | WikipediaGS | F1 (%) | 67 | TURL |
| Data Integration | WikiTables-TURL-CPA | F1 (%) | 94.91 | TURL |
| Table annotation | T2Dv2 | Accuracy (%) | 96.2 | TURL |
| Table annotation | WikiTables-TURL-CTA | F1 (%) | 94.75 | TURL |
| Table annotation | WikipediaGS-CTA | Accuracy (%) | 74.6 | TURL |
| Table annotation | WikiTables-TURL-CEA | F1 (%) | 68 | TURL |
| Table annotation | WikipediaGS | F1 (%) | 67 | TURL |
| Table annotation | WikiTables-TURL-CPA | F1 (%) | 94.91 | TURL |