Skip to main navigation Skip to search Skip to main content

Pretrained Knowledge Base Embeddings for improved Sentential Relation Extraction

Andrea Papaluca, Daniel Krefl, Hanna Suominen, Artem Lenskiy

Research output: Chapter in Book/Report/Conference proceedingConference Paperpeer-review

10 Citations (Scopus)

Abstract

In this work we put forward to combine pre-trained knowledge base graph embeddings with transformer based language models to improve performance on the sentential Relation Extraction task in natural language processing. Our proposed model is based on a simple variation of existing models to incorporate off-task pre-trained graph embeddings with an on-task finetuned BERT encoder. We perform a detailed statistical evaluation of the model on standard datasets. We provide evidence that the added graph embeddings improve the performance, making such a simple approach competitive with the state-of-the-art models that perform explicit on-task training of the graph embeddings. Furthermore, we observe for the underlying BERT model an interesting power-law scaling behavior between the variance of the F1 score obtained for a relation class and its support in terms of training examples.

Original languageEnglish
Title of host publicationACL 2022 - 60th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Student Research Workshop
EditorsSamuel Louvan, Andrea Madotto, Brielen Madureira
PublisherAssociation for Computational Linguistics (ACL)
Pages373-382
Number of pages10
ISBN (Electronic)9781955917230
Publication statusPublished - 2022
Event60th Annual Meeting of the Association for Computational Linguistics, ACL 2022 - Dublin, Ireland
Duration: 22 May 202227 May 2022

Publication series

NameProceedings of the Annual Meeting of the Association for Computational Linguistics
ISSN (Print)0736-587X

Conference

Conference60th Annual Meeting of the Association for Computational Linguistics, ACL 2022
Country/TerritoryIreland
CityDublin
Period22/05/2227/05/22

Fingerprint

Dive into the research topics of 'Pretrained Knowledge Base Embeddings for improved Sentential Relation Extraction'. Together they form a unique fingerprint.

Cite this