SOTAVerified

Optimal Embedding Guided Negative Sample Generation for Knowledge Graph Link Prediction

2025-04-04Code Available0· sign in to hype

Makoto Takamoto, Daniel Oñoro-Rubio, Wiem Ben Rim, Takashi Maruyama, Bhushan Kotnis

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Knowledge graph embedding (KGE) models encode the structural information of knowledge graphs to predicting new links. Effective training of these models requires distinguishing between positive and negative samples with high precision. Although prior research has shown that improving the quality of negative samples can significantly enhance model accuracy, identifying high-quality negative samples remains a challenging problem. This paper theoretically investigates the condition under which negative samples lead to optimal KG embedding and identifies a sufficient condition for an effective negative sample distribution. Based on this theoretical foundation, we propose Embedding MUtation (EMU), a novel framework that generates negative samples satisfying this condition, in contrast to conventional methods that focus on identifying challenging negative samples within the training data. Importantly, the simplicity of EMU ensures seamless integration with existing KGE models and negative sampling methods. To evaluate its efficacy, we conducted comprehensive experiments across multiple datasets. The results consistently demonstrate significant improvements in link prediction performance across various KGE models and negative sampling methods. Notably, EMU enables performance improvements comparable to those achieved by models with embedding dimension five times larger. An implementation of the method and experiments are available at https://github.com/nec-research/EMU-KG.

Tasks

Reproductions