Abstract
Infusing factual knowledge into pretrained models is fundamental for many knowledge-intensive tasks. In this paper, we propose Mixture-of-Partitions (MoP), an infusion approach that can handle a very large knowledge graph (KG) by partitioning it into smaller sub-graphs and infusing their specific knowledge into various BERT models using lightweight adapters. To leverage the overall factual knowledge for a target task, these sub-graph adapters are further fine-tuned along with the underlying BERT through a mixture layer. We evaluate our MoP with three biomedical BERTs (SciBERT, BioBERT, PubmedBERT) on six downstream tasks (inc. NLI, QA, Classification), and the results show that our MoP consistently enhances the underlying BERTs in task performance, and achieves new SOTA performances on five evaluated datasets.
Original language | English |
---|---|
Title of host publication | Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing |
Editors | Xuanjing Huang, Lucia Specia, Scott Wen-tau Yih |
Place of Publication | Stroudsburg PA USA |
Publisher | Association for Computational Linguistics (ACL) |
Pages | 4672–4681 |
Number of pages | 10 |
ISBN (Electronic) | 9781955917094 |
Publication status | Published - 2021 |
Event | Empirical Methods in Natural Language Processing 2021 - Online, Punta Cana, Dominican Republic Duration: 7 Nov 2021 → 11 Nov 2021 https://2021.emnlp.org/ (Website) https://aclanthology.org/2021.emnlp-main.0/ (Proceedings) https://aclanthology.org/2021.findings-emnlp.0/ (Proceedings - findings) |
Conference
Conference | Empirical Methods in Natural Language Processing 2021 |
---|---|
Abbreviated title | EMNLP 2021 |
Country/Territory | Dominican Republic |
City | Punta Cana |
Period | 7/11/21 → 11/11/21 |
Internet address |
|