• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Scopus
  4. Language Model Guided Knowledge Graph Embeddings
 
  • Details
  • Full
Options
2022
Journal Article
Title

Language Model Guided Knowledge Graph Embeddings

Abstract
Knowledge graph embedding models have become a popular approach for knowledge graph completion through predicting the plausibility of (potential) triples. This is performed by transforming the entities and relations of the knowledge graph into an embedding space. However, knowledge graphs often include further textual information stored in literal, which is ignored by such embedding models. As a consequence, the learning process stays limited to the structure and the connections between the entities, which has the potential to negatively influence the performance. We bridge this gap by leveraging the capabilities of pre-trained language models to include textual knowledge in the learning process of embedding models. This is achieved by introducing a new loss function that guides embedding models in measuring the likelihood of triples by taking such complementary knowledge into consideration. The proposed solution is a model-independent loss function that can be plugged into any knowledge graph embedding model. In this paper, Sentence-BERT and fastText are used as pre-trained language models from which the embeddings of the textual knowledge are obtained and injected into the loss function. The loss function contains a trainable slack variable that determines the degree to which the language models influence the plausibility of triples. Our experimental evaluation on six benchmarks, namely Nations, UMLS, WordNet, and three versions of CodEx confirms the advantage of using pre-trained language models for boosting the accuracy of knowledge graph embedding models. We showcase this by performing evaluations on top of the five well-known knowledge graph embedding models such as TransE, RotatE, ComplEx, DistMult, and QuatE. The results show an improvement in accuracy up to 9% on UMLS dataset for the Distmult model and 4.2% on the Nations dataset for the ComplEx model when they are guided by pre-trained language models. We additionally studied the effect of multiple factors such as the structure of the knowledge graphs and training steps and presented them as ablation studies.
Author(s)
Alam, Mirza Mohtashim
Institute for Applied Informatics (InfAI)
Rony, Md Rashad Al Hasan
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Nayyeri, Mojtaba
Universität Bonn
Mohiuddin, Karishma
Universität Bonn
Akter, M.S.T.M.
Universität Bonn
Vahdati, Sahar
Institute for Applied Informatics (InfAI)
Lehmann, Jens
Institute for Applied Informatics (InfAI)
Journal
IEEE access  
Project(s)
Foundations of Trustworthy AI - Integrating Reasoning, Learning and Optimization  
Copernicus Artificial Intelligence Services and data fusion with other distributed data sources and processing at the edge to support DIAS and HPC infrastructures  
01IS18050F  
Aufbau einer führenden Sprachassistenzplattform "Made in Germany"  
Funder
European Commission
Deutsches Bundesministerium für Bildung und Forschung  
Bundesministerium für Wirtschaft und Klimaschutz -BMWK-
Open Access
DOI
10.1109/ACCESS.2022.3191666
Language
English
Fraunhofer-Institut für Intelligente Analyse- und Informationssysteme IAIS  
Keyword(s)
  • Knowledge graph

  • knowledge graph embeddings

  • language models

  • link prediction

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024