PK-BERT: Knowledge Enhanced Pre-trained Models with Prompt for Few-Shot Learning

研究成果: Conference contribution同行評審

2 引文 斯高帕斯(Scopus)

摘要

The amount of data in some fields are scarce because they are difficult or expensive to obtain. The general practice is to pre-train a model on similar data sets and fine-tune the models in downstream tasks by transfer learning. The pre-trained models could learn the general language representation from large-scale corpora but their downstream task may be different from the pre-trained tasks in form and type. It also lacks related semantic knowledge. Therefore, we propose PK-BERT—Knowledge Enhanced Pre-trained Models with Prompt for Few-shot Learning. It (1) achieves few-shot learning by using small samples with pre-trained models; (2) constructs the prefix that contains the masked label to shorten the gap between downstream task and pre-trained task; (3) uses the explicit representation to inject knowledge graph triples into the text to enhance the sentence information; and (4) uses masked language modelling (MLM) head to convert the classification task into generation task. The experiments show that our proposed model PK-BERT achieves better results.

原文English
主出版物標題Computer and Information Science
編輯Roger Lee
發行者Springer Science and Business Media Deutschland GmbH
頁面31-44
頁數14
ISBN(列印)9783031121265
DOIs
出版狀態Published - 2023
事件22nd IEEE/ACIS International Conference on Computer and Information Science, ICIS 2022 - Zhuhai, China
持續時間: 26 6月 202228 6月 2022

出版系列

名字Studies in Computational Intelligence
1055
ISSN(列印)1860-949X
ISSN(電子)1860-9503

Conference

Conference22nd IEEE/ACIS International Conference on Computer and Information Science, ICIS 2022
國家/地區China
城市Zhuhai
期間26/06/2228/06/22

指紋

深入研究「PK-BERT: Knowledge Enhanced Pre-trained Models with Prompt for Few-Shot Learning」主題。共同形成了獨特的指紋。

引用此