Cristina Cornelio, Veronika Thost
Logical rules are a popular knowledge representation language in many domains, representing background knowledge and encoding information that can be derived from given facts in a compact form. However, rule formulation is a complex process that requires deep domain expertise,and is further challenged by today's often large, heterogeneous, and incomplete knowledge graphs. Several approaches for learning rules automatically, given a set of input example facts,have been proposed over time, including, more recently, neural systems. Yet, the area is missing adequate datasets and evaluation approaches: existing datasets often resemble toy examples that neither cover the various kinds of dependencies between rules nor allow for testing scalability. We present a tool for generating different kinds of datasets and for evaluating rule learning systems, including new performance measures.
| Task | Dataset | Metric | Value | Model |
|---|---|---|---|---|
| Inductive logic programming | RuDaS | H-Score | 0.2321 | AMIE+ |
| Inductive logic programming | RuDaS | R-Score | 0.335 | AMIE+ |
| Inductive logic programming | RuDaS | H-Score | 0.152 | FOIL |
| Inductive logic programming | RuDaS | R-Score | 0.2728 | FOIL |
| Inductive logic programming | RuDaS | H-Score | 0.1025 | Neural-LP |
| Inductive logic programming | RuDaS | R-Score | 0.1906 | Neural-LP |
| Inductive logic programming | RuDaS | H-Score | 0.0728 | NTP |
| Inductive logic programming | RuDaS | R-Score | 0.1811 | NTP |