RuDaS: Synthetic Datasets for Rule Learning and Evaluation Tools
Cristina Cornelio, Veronika Thost
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/IBM/RuDaSOfficialIn papernone★ 0
Abstract
Logical rules are a popular knowledge representation language in many domains, representing background knowledge and encoding information that can be derived from given facts in a compact form. However, rule formulation is a complex process that requires deep domain expertise,and is further challenged by today's often large, heterogeneous, and incomplete knowledge graphs. Several approaches for learning rules automatically, given a set of input example facts,have been proposed over time, including, more recently, neural systems. Yet, the area is missing adequate datasets and evaluation approaches: existing datasets often resemble toy examples that neither cover the various kinds of dependencies between rules nor allow for testing scalability. We present a tool for generating different kinds of datasets and for evaluating rule learning systems, including new performance measures.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| RuDaS | AMIE+ | H-Score | 0.23 | — | Unverified |
| RuDaS | FOIL | H-Score | 0.15 | — | Unverified |
| RuDaS | Neural-LP | H-Score | 0.1 | — | Unverified |
| RuDaS | NTP | H-Score | 0.07 | — | Unverified |