Repository logo
 

Mixture-of-Partitions: Infusing Large Biomedical Knowledge Graphs into BERT

Accepted version
Peer-reviewed

Type

Conference Object

Change log

Authors

Clark, TH 
Shareghi, E 

Abstract

Infusing factual knowledge into pre-trained models is fundamental for many knowledge-intensive tasks. In this paper, we proposed Mixture-of-Partitions (MoP), an infusion approach that can handle a very large knowledge graph (KG) by partitioning it into smaller sub-graphs and infusing their specific knowledge into various BERT models using lightweight adapters. To leverage the overall factual knowledge for a target task, these sub-graph adapters are further fine-tuned along with the underlying BERT through a mixture layer. We evaluate our MoP with three biomedical BERTs (SciBERT, BioBERT, PubmedBERT) on six downstream tasks (inc. NLI, QA, Classification), and the results show that our MoP consistently enhances the underlying BERTs in task performance, and achieves new SOTA performances on five evaluated datasets.

Description

Keywords

cs.CL, cs.CL

Journal Title

EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings

Conference Name

The 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP 2021)

Journal ISSN

Volume Title

Publisher

Rights

All rights reserved
Sponsorship
ESRC (ES/T012277/1)