SOTAVerified

The Curious Case of Stacking Boosted Relational Dependency Networks

2020-10-19NeurIPS Workshop ICBINB 2020Unverified0· sign in to hype

Siwen Yan, Devendra Singh Dhami, Sriraam Natarajan

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Reducing bias while learning and inference is an important requirement to achieve generalizable and better performing models. The method of stacking took the first step towards creating such models by reducing inference bias but the question of combining stacking with a model that reduces learning bias is still largely unanswered. In statistical relational learning, ensemble models of relational trees such as boosted relational dependency networks (RDN-Boost) are shown to reduce the learning bias. We combine RDN-Boost and stacking methods with the aim of reducing both learning and inference bias subsequently resulting in better overall performance. However, our evaluation on three relational data sets shows no significant performance improvement over the baseline models.

Tasks

Reproductions