Markus Eberts, Adrian Ulges
We introduce SpERT, an attention model for span-based joint entity and relation extraction. Our key contribution is a light-weight reasoning on BERT embeddings, which features entity recognition and filtering, as well as relation classification with a localized, marker-free context representation. The model is trained using strong within-sentence negative samples, which are efficiently extracted in a single BERT pass. These aspects facilitate a search over all spans in the sentence. In ablation studies, we demonstrate the benefits of pre-training, strong negative sampling and localized context. Our model outperforms prior work by up to 2.6% F1 score on several datasets for joint entity and relation extraction.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Relation Extraction | Adverse Drug Events (ADE) Corpus | NER Macro F1 | 89.25 | SpERT (without overlap) |
| Relation Extraction | Adverse Drug Events (ADE) Corpus | RE+ Macro F1 | 79.24 | SpERT (without overlap) |
| Relation Extraction | Adverse Drug Events (ADE) Corpus | NER Macro F1 | 89.28 | SpERT (with overlap) |
| Relation Extraction | Adverse Drug Events (ADE) Corpus | RE+ Macro F1 | 78.84 | SpERT (with overlap) |
| Relation Extraction | CoNLL04 | NER Macro F1 | 86.25 | SpERT |
| Relation Extraction | CoNLL04 | NER Micro F1 | 88.94 | SpERT |
| Relation Extraction | CoNLL04 | RE+ Macro F1 | 72.87 | SpERT |
| Relation Extraction | CoNLL04 | RE+ Micro F1 | 71.47 | SpERT |
| Relation Extraction | SciERC | Entity F1 | 70.3 | SpERT (with overlap) |
| Relation Extraction | SciERC | Relation F1 | 50.84 | SpERT (with overlap) |
| Relation Extraction | SciERC | Entity F1 | 70.33 | SpERT |
| Information Extraction | SciERC | Entity F1 | 70.3 | SpERT (with overlap) |
| Information Extraction | SciERC | Relation F1 | 50.84 | SpERT (with overlap) |
| Information Extraction | SciERC | Entity F1 | 70.33 | SpERT |
| Named Entity Recognition (NER) | SciERC | F1 | 70.33 | SpERT |