TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Robust (Controlled) Table-to-Text Generation with Structur...

Robust (Controlled) Table-to-Text Generation with Structure-Aware Equivariance Learning

Fei Wang, Zhewei Xu, Pedro Szekely, Muhao Chen

2022-05-08NAACL 2022 7Data-to-Text GenerationText GenerationData AugmentationTable-to-Text GenerationGraph Attention
PaperPDFCode(official)

Abstract

Controlled table-to-text generation seeks to generate natural language descriptions for highlighted subparts of a table. Previous SOTA systems still employ a sequence-to-sequence generation method, which merely captures the table as a linear structure and is brittle when table layouts change. We seek to go beyond this paradigm by (1) effectively expressing the relations of content pieces in the table, and (2) making our model robust to content-invariant structural transformations. Accordingly, we propose an equivariance learning framework, which encodes tables with a structure-aware self-attention mechanism. This prunes the full self-attention structure into an order-invariant graph attention that captures the connected graph structure of cells belonging to the same row or column, and it differentiates between relevant cells and irrelevant cells from the structural perspective. Our framework also modifies the positional encoding mechanism to preserve the relative position of tokens in the same cell but enforce position invariance among different cells. Our technology is free to be plugged into existing table-to-text generation models, and has improved T5-based models to offer better performance on ToTTo and HiTab. Moreover, on a harder version of ToTTo, we preserve promising performance, while previous SOTA systems, even with transformation-based data augmentation, have seen significant performance drops. Our code is available at https://github.com/luka-group/Lattice.

Results

TaskDatasetMetricValueModel
Text GenerationToTToBLEU48.4LATTICE (T5-base)
Text GenerationToTToPARENT58.1LATTICE (T5-base)
Data-to-Text GenerationToTToBLEU48.4LATTICE (T5-base)
Data-to-Text GenerationToTToPARENT58.1LATTICE (T5-base)

Related Papers

Making Language Model a Hierarchical Classifier and Generator2025-07-17Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17Pixel Perfect MegaMed: A Megapixel-Scale Vision-Language Foundation Model for Generating High Resolution Medical Images2025-07-17Mitigating Object Hallucinations via Sentence-Level Early Intervention2025-07-16Similarity-Guided Diffusion for Contrastive Sequential Recommendation2025-07-16Catching Bid-rigging Cartels with Graph Attention Neural Networks2025-07-16The Devil behind the mask: An emergent safety vulnerability of Diffusion LLMs2025-07-15Seq vs Seq: An Open Suite of Paired Encoders and Decoders2025-07-15