RotatE: Knowledge Graph Embedding by Relational Rotation in Complex SpaceDownload PDF

Published: 21 Dec 2018, Last Modified: 22 Oct 2023ICLR 2019 Conference Blind SubmissionReaders: Everyone
Abstract: We study the problem of learning representations of entities and relations in knowledge graphs for predicting missing links. The success of such a task heavily relies on the ability of modeling and inferring the patterns of (or between) the relations. In this paper, we present a new approach for knowledge graph embedding called RotatE, which is able to model and infer various relation patterns including: symmetry/antisymmetry, inversion, and composition. Specifically, the RotatE model defines each relation as a rotation from the source entity to the target entity in the complex vector space. In addition, we propose a novel self-adversarial negative sampling technique for efficiently and effectively training the RotatE model. Experimental results on multiple benchmark knowledge graphs show that the proposed RotatE model is not only scalable, but also able to infer and model various relation patterns and significantly outperform existing state-of-the-art models for link prediction.
Keywords: knowledge graph embedding, knowledge graph completion, adversarial sampling
TL;DR: A new state-of-the-art approach for knowledge graph embedding.
Code: [![github](/images/github_icon.svg) DeepGraphLearning/KnowledgeGraphEmbedding](https://github.com/DeepGraphLearning/KnowledgeGraphEmbedding) + [![Papers with Code](/images/pwc_icon.svg) 6 community implementations](https://paperswithcode.com/paper/?openreview=HkgEQnRqYQ)
Data: [FB122](https://paperswithcode.com/dataset/fb122), [FB15k](https://paperswithcode.com/dataset/fb15k), [FB15k-237](https://paperswithcode.com/dataset/fb15k-237), [OGB](https://paperswithcode.com/dataset/ogb), [WN18](https://paperswithcode.com/dataset/wn18), [WN18RR](https://paperswithcode.com/dataset/wn18rr)
Community Implementations: [![CatalyzeX](/images/catalyzex_icon.svg) 2 code implementations](https://www.catalyzex.com/paper/arxiv:1902.10197/code)
41 Replies

Loading