Ningyu Zhang, Xiang Chen, Xin Xie, Shumin Deng, Chuanqi Tan, Mosha Chen, Fei Huang, Luo Si, Huajun Chen
Document-level relation extraction aims to extract relations among multiple entity pairs from a document. Previously proposed graph-based or transformer-based models utilize the entities independently, regardless of global information among relational triples. This paper approaches the problem by predicting an entity-level relation matrix to capture local and global information, parallel to the semantic segmentation task in computer vision. Herein, we propose a Document U-shaped Network for document-level relation extraction. Specifically, we leverage an encoder module to capture the context information of entities and a U-shaped segmentation module over the image-style feature map to capture global interdependency among triples. Experimental results show that our approach can obtain state-of-the-art performance on three benchmark datasets DocRED, CDR, and GDA.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Relation Extraction | DocRED | F1 | 64.55 | DocuNet-RoBERTa-large |
| Relation Extraction | DocRED | Ign F1 | 62.4 | DocuNet-RoBERTa-large |
| Relation Extraction | ReDocRED | F1 | 77.87 | DocuNET |
| Relation Extraction | ReDocRED | Ign F1 | 77.26 | DocuNET |
| Relation Extraction | GDA | F1 | 85.3 | DocuNet-SciBERTbase |
| Relation Extraction | CDR | F1 | 76.3 | DocuNet-SciBERTbase |