Přehled o publikaci
2021
On Training Knowledge Graph Embedding Models
MOHAMED, Sameh K; Vít NOVACEK and Emir MUNOZBasic information
Original name
On Training Knowledge Graph Embedding Models
Authors
MOHAMED, Sameh K; Vít NOVACEK and Emir MUNOZ
Edition
Information, Switzerland, MDPI, 2021, 2078-2489
Other information
Language
English
Type of outcome
Article in a journal
Country of publisher
Switzerland
Confidentiality degree
is not subject to a state or trade secret
References:
Marked to be transferred to RIV
Yes
RIV identification code
RIV/00216224:14330/21:00121336
Organization
Fakulta informatiky – Repository – Repository
UT WoS
EID Scopus
Keywords in English
loss functions; knowledge graph embeddings; link prediction
Changed: 24/5/2022 05:21, RNDr. Daniel Jakubík
Abstract
In the original language
Knowledge graph embedding (KGE) models have become popular means for making discoveries in knowledge graphs (e.g., RDF graphs) in an efficient and scalable manner. The key to success of these models is their ability to learn low-rank vector representations for knowledge graph entities and relations. Despite the rapid development of KGE models, state-of-the-art approaches have mostly focused on new ways to represent embeddings interaction functions (i.e., scoring functions). In this paper, we argue that the choice of other training components such as the loss function, hyperparameters and negative sampling strategies can also have substantial impact on the model efficiency. This area has been rather neglected by previous works so far and our contribution is towards closing this gap by a thorough analysis of possible choices of training loss functions, hyperparameters and negative sampling techniques. We finally investigate the effects of specific choices on the scalability and accuracy of knowledge graph embedding models.