Yufang Hou, Charles Jochim, Martin Gleize, Francesca Bonin, Debasis Ganguly
While the fast-paced inception of novel tasks and new datasets helps foster active research in a community towards interesting directions, keeping track of the abundance of research activity in different areas on different datasets is likely to become increasingly difficult. The community could greatly benefit from an automatic system able to summarize scientific results, e.g., in the form of a leaderboard. In this paper we build two datasets and develop a framework (TDMS-IE) aimed at automatically extracting task, dataset, metric and score from NLP papers, towards the automatic construction of leaderboards. Experiments show that our model outperforms several baselines by a large margin. Our model is a first step towards automatic leaderboard construction, e.g., in the NLP domain.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Information Retrieval | NLP-TDMS (Exp, arXiv only) | Macro F1 | 8.8 | TDMS-IE |
| Information Retrieval | NLP-TDMS (Exp, arXiv only) | Macro Precision | 9.5 | TDMS-IE |
| Information Retrieval | NLP-TDMS (Exp, arXiv only) | Macro Recall | 8.6 | TDMS-IE |
| Information Retrieval | NLP-TDMS (Exp, arXiv only) | Micro F1 | 7.5 | TDMS-IE |
| Information Retrieval | NLP-TDMS (Exp, arXiv only) | Micro Precision | 6.8 | TDMS-IE |
| Information Retrieval | NLP-TDMS (Exp, arXiv only) | Micro Recall | 8.4 | TDMS-IE |