期刊文献+

基于提示学习的小样本文本分类方法 被引量:4

Few-shot text classification method based on prompt learning
下载PDF
导出
摘要 文本分类任务通常依赖足量的标注数据,针对低资源场景下的分类模型在小样本上的过拟合问题,提出一种基于提示学习的小样本文本分类方法BERT-P-Tuning。首先,利用预训练模型BERT(Bidirectional Encoder Representations from Transformers)在标注样本上学习到最优的提示模板;然后,在每条样本中补充提示模板和空缺,将文本分类任务转化为完形填空任务;最后,通过预测空缺位置概率最高的词并结合它与标签之间的映射关系得到最终的标签。在公开数据集FewCLUE上的短文本分类任务上进行实验,实验结果表明,所提方法相较于基于BERT微调的方法在评价指标上有显著提高。所提方法在二分类任务上的准确率与F1值分别提升了25.2和26.7个百分点,在多分类任务上的准确率与F1值分别提升了6.6和8.0个百分点。相较于手动构建模板的PET(Pattern Exploiting Training)方法,所提方法在两个任务上的准确率分别提升了2.9和2.8个百分点,F1值分别提升了4.4和4.2个百分点,验证了预训练模型应用在小样本任务的有效性。 Text classification tasks usually rely on sufficient labeled data.Concerning the over-fitting problem of classification models on samples with small size in low resource scenarios,a few-shot text classification method based on prompt learning called BERT-P-Tuning was proposed.Firstly,the pre-trained model BERT(Bidirectional Encoder Representations from Transformers)was used to learn the optimal prompt template from labeled samples.Then,the prompt template and vacancy were filled in each sample,and the text classification task was transformed into the cloze test task.Finally,the final labels were obtained by predicting the word with the highest probability of the vacant positions and combining the mapping relationship between it and labels.Experimental results on the short text classification tasks of public dataset FewCLUE show that the proposed method have significantly improved the evaluation indicators compared to the BERT fine-tuning based method.In specific,the proposed method has the accuracy and F1 score increased by 25.2 and 26.7 percentage points respectively on the binary classification task,and the proposed method has the accuracy and F1 score increased by 6.6 and 8.0 percentage points respectively on the multi-class classification task.Compared with the PET(Pattern Exploiting Training)method of constructing templates manually,the proposed method has the accuracy increased by 2.9 and 2.8 percentage points respectively on two tasks,and the F1 score increased by 4.4 and 4.2 percentage points respectively on two tasks.The above verifies the effectiveness of applying pre-trained model on few-shot tasks.
作者 于碧辉 蔡兴业 魏靖烜 YU Bihui;CAI Xingye;WEI Jingxuan(University of Chinese Academy of Sciences,Beijing 100049,China;Shenyang Institute of Computing Technology,Chinese Academy of Sciences,Shenyang Liaoning 110168,China)
出处 《计算机应用》 CSCD 北大核心 2023年第9期2735-2740,共6页 journal of Computer Applications
基金 国家重点研发计划项目(2019YFB1405803)。
关键词 小样本学习 文本分类 预训练模型 提示学习 自适应模板 few-shot learning text classification pre-trained model prompt learning adaptive template
  • 相关文献

参考文献3

二级参考文献34

共引文献311

同被引文献33

引证文献4

二级引证文献4

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部