Knowledge Graph Entity Alignment Using Relation Structural Similarity

Knowledge Graph Entity Alignment Using Relation Structural Similarity

Yanhui Peng, Jing Zhang, Cangqi Zhou, Shunmei Meng
Copyright: © 2022 |Pages: 19
DOI: 10.4018/JDM.305733
OnDemand:
(Individual Articles)
Available
$37.50
No Current Special Offers
TOTAL SAVINGS: $37.50

Abstract

Embedding-based entity alignment, which represents knowledge graphs as low-dimensional embeddings and finds entities in different knowledge graphs that semantically represent the same real-world entity by measuring the similarities between entity embeddings, has achieved promising results. However, existing methods are still challenged by the error accumulation of embeddings along multi-step paths and the semantic information loss. This paper proposes a novel embedding-based entity alignment method that iteratively aligns both entities and relations with high similarities as training data. Newly-aligned entities and relations are used to calibrate the corresponding embeddings in the unified embedding space, which reduces the error accumulation. To reduce the negative impact of semantic information loss, the authors propose to use relation structural similarity instead of embedding similarity to align relations. Experimental results on five widely used real-world datasets show that the proposed method significantly outperforms several state-of-the-art methods for entity alignment.
Article Preview
Top

Introduction

Knowledge graphs, such as WordNet (Miller, 1995), Freebase (Bollacker et al., 2008), and DBpedia (Lehmann et al., 2015), store a great number of knowledge triples in the form of directed graphs, where each node denotes an entity, and each edge denotes a relation. A knowledge triple (head entity, relation, tail entity) (denoted by (, , ) in this paper) stands for an edge with two end nodes in the graph, indicating that there exists a specific relationship between the head and tail entities. Recently, applications of knowledge graphs in artificial intelligence domains, including information systems (Saghafi & Wand, 2020), recommender systems (Zhang et al., 2016), information retrieval (Xiong et al., 2017), question-answering systems (Hao et al., 2017; Huang et al., 2019), and natural language processing (Yang & Mitchell, 2019), have gained more and more attention. There often exist multiple knowledge graphs in overlapped application domains or even the same domain. Moreover, some knowledge graphs are created with crowdsourcing (Chai et al., 2018), due to the uncertainty of crowdsourcing annotation (Zhang, Wu, & Sheng, 2016), they are generally incomplete, resulting in difficulties for a single knowledge graph to meet the various requirements of many AI applications. Integrating heterogeneous knowledge among different knowledge graphs through entity alignment can make knowledge graphs satisfy wider requirements in different applications. However, this goal is not easy to achieve. Because entities and relations in different knowledge graphs may be represented by different languages or symbols, it is generally ineffective for symbolic-based methods (Ngomo & Auer, 2011; Volz et al., 2009) to precisely align entities. For example, “https://www.wikidata.org/entity/ Q7769108” in Wikidata are the same entity in the real world, it is almost impossible to measure the similarity between them by symbolic-based methods. In order to solve this difficult problem, researchers have proposed quite a few embedding-based entity alignment approaches. Such approaches first embed symbolic entities and relations of a knowledge graph into a low-dimensional vector space and then find out entity alignments by measuring the similarities between the corresponding embeddings.

Generally, knowledge graph embedding models focus on representing a single knowledge graph as vectors to capture the semantics of entities and relations according to their structural information in the knowledge graph. For example, as one of the most popular knowledge graph embedding models in entity alignment methods, TransE (Bordes et al., 2013) interprets each relation as a translation from the head entity to the tail entity. However, such separately-trained embeddings cannot be directly used to compute the similarity between entities from different knowledge graphs, because embeddings of different knowledge graphs are in different semantic spaces (i.e., vector space).

Complete Article List

Search this Journal:
Reset
Volume 35: 1 Issue (2024)
Volume 34: 3 Issues (2023)
Volume 33: 5 Issues (2022): 4 Released, 1 Forthcoming
Volume 32: 4 Issues (2021)
Volume 31: 4 Issues (2020)
Volume 30: 4 Issues (2019)
Volume 29: 4 Issues (2018)
Volume 28: 4 Issues (2017)
Volume 27: 4 Issues (2016)
Volume 26: 4 Issues (2015)
Volume 25: 4 Issues (2014)
Volume 24: 4 Issues (2013)
Volume 23: 4 Issues (2012)
Volume 22: 4 Issues (2011)
Volume 21: 4 Issues (2010)
Volume 20: 4 Issues (2009)
Volume 19: 4 Issues (2008)
Volume 18: 4 Issues (2007)
Volume 17: 4 Issues (2006)
Volume 16: 4 Issues (2005)
Volume 15: 4 Issues (2004)
Volume 14: 4 Issues (2003)
Volume 13: 4 Issues (2002)
Volume 12: 4 Issues (2001)
Volume 11: 4 Issues (2000)
Volume 10: 4 Issues (1999)
Volume 9: 4 Issues (1998)
Volume 8: 4 Issues (1997)
Volume 7: 4 Issues (1996)
Volume 6: 4 Issues (1995)
Volume 5: 4 Issues (1994)
Volume 4: 4 Issues (1993)
Volume 3: 4 Issues (1992)
Volume 2: 4 Issues (1991)
Volume 1: 2 Issues (1990)
View Complete Journal Contents Listing