Contrastive Language-Entity Pre-training for Richer Knowledge Graph Embedding

Andrea Papaluca*, Daniel Krefl, Artem Lensky, Hanna Suominen

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In this work we propose a pretraining procedure that aligns a graph encoder and a text encoder to learn a common multi-modal graph-text embedding space. The alignment is obtained by training a model to predict the correct associations between Knowledge Graph nodes and their corresponding descriptions. We test the procedure with two popular Knowledge Bases: Wikidata (formerly Freebase) and YAGO. Our results indicate that such a pretraining method allows for link prediction without the need for additional fine-tuning. Furthermore, we demonstrate that a graph encoder pretrained on the description matching task allows for improved link prediction performance after fine-tuning, without the need for providing node descriptions as additional inputs. We make available the code used in the experiments on GitHub(https://github.com/BrunoLiegiBastonLiegi/CLEP) under the MIT license to encourage further work.

Original languageEnglish
Title of host publicationPattern Recognition and Artificial Intelligence - 4th International Conference, ICPRAI 2024, Proceedings
EditorsChristian Wallraven, Cheng-Lin Liu, Arun Ross
PublisherSpringer Science and Business Media Deutschland GmbH
Pages233-246
Number of pages14
ISBN (Print)9789819787012
DOIs
Publication statusPublished - 2025
Event4th International Conference on Pattern Recognition and Artificial Intelligence, ICPRAI 2024 - Jeju Island, Korea, Republic of
Duration: 3 Jul 20246 Jul 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14892 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference4th International Conference on Pattern Recognition and Artificial Intelligence, ICPRAI 2024
Country/TerritoryKorea, Republic of
CityJeju Island
Period3/07/246/07/24

Fingerprint

Dive into the research topics of 'Contrastive Language-Entity Pre-training for Richer Knowledge Graph Embedding'. Together they form a unique fingerprint.

Cite this