Representation Learning of Knowledge Graphs with Entity Descriptions
Citations Over TimeTop 1% of 2016 papers
Abstract
Representation learning (RL) of knowledge graphs aims to project both entities and relations into a continuous low-dimensional space. Most methods concentrate on learning representations with knowledge triples indicating relations between entities. In fact, in most knowledge graphs there are usually concise descriptions for entities, which cannot be well utilized by existing methods. In this paper, we propose a novel RL method for knowledge graphs taking advantages of entity descriptions. More specifically, we explore two encoders, including continuous bag-of-words and deep convolutional neural models to encode semantics of entity descriptions. We further learn knowledge representations with both triples and descriptions. We evaluate our method on two tasks, including knowledge graph completion and entity classification. Experimental results on real-world datasets show that, our method outperforms other baselines on the two tasks, especially under the zero-shot setting, which indicates that our method is capable of building representations for novel entities according to their descriptions. The source code of this paper can be obtained from https://github.com/xrb92/DKRL.
Related Papers
- → Be Concise and Precise: Synthesizing Open-Domain Entity Descriptions from Facts(2019)4 cited
- → Review of Biomedical Knowledge and Data Representation with Conceptual Graphs(1998)23 cited
- → Cross-Lingual Entity Query from Large-Scale Knowledge Graphs(2015)3 cited
- → DB-MAT: Knowledge acquisition, processing and NL generation using conceptual graphs(1996)13 cited
- → Which Is the Tallest Building in Europe? Representing and Reasoning About Knowledge(2019)