Papers
arxiv:2002.01808

K-Adapter: Infusing Knowledge into Pre-Trained Models with Adapters

Published on Feb 5, 2020
Authors:
,
,
,
,
,
,
,
,

Abstract

We study the problem of injecting knowledge into large pre-trained models like BERT and Ro<PRE_TAG>BERTa</POST_TAG>. Existing methods typically update the original parameters of pre-trained models when injecting knowledge. However, when multiple kinds of knowledge are injected, the historically injected knowledge would be flushed away. To address this, we propose K-Adapter, a framework that retains the original parameters of the pre-trained model fixed and supports the development of versatile knowledge-infused model. Taking Ro<PRE_TAG>BERTa</POST_TAG> as the backbone model, K-Adapter has a neural adapter for each kind of infused knowledge, like a plug-in connected to Ro<PRE_TAG>BERTa</POST_TAG>. There is no information flow between different adapters, thus multiple adapters can be efficiently trained in a distributed way. As a case study, we inject two kinds of knowledge in this work, including (1) factual knowledge obtained from automatically aligned text-triplets on Wikipedia and Wikidata and (2) linguistic knowledge obtained via dependency parsing. Results on three knowledge-driven tasks, including relation classification, entity typing, and question answering, demonstrate that each adapter improves the performance and the combination of both adapters brings further improvements. Further analysis indicates that K-Adapter captures versatile knowledge than Ro<PRE_TAG>BERTa</POST_TAG>.

Community

Sign up or log in to comment

Models citing this paper 3

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2002.01808 in a dataset README.md to link it from this page.

Spaces citing this paper 2

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.