期刊文献+

基于BERT-BiLSTM模型的短文本自动评分系统 被引量:7

Short text automatic scoring system based on BERT-BiLSTM model
下载PDF
导出
摘要 针对短文本自动评分中存在的特征稀疏、一词多义及上下文关联信息少等问题,提出一种基于BERT-BiLSTM(bidirectional encoder representations from transformers-bidirectional long short-term memory)的短文本自动评分模型.使用BERT(bidirectional encoder representations from transformers)语言模型预训练大规模语料库习得通用语言的语义特征,通过预训练好的BERT语言模型预微调下游具体任务的短文本数据集习得短文本的语义特征和关键词特定含义,再通过BiLSTM(bidirectional long short-term memory)捕获深层次上下文关联信息,最后将获得的特征向量输入Softmax回归模型进行自动评分.实验结果表明,对比CNN(convolutional neural networks)、CharCNN(character-level CNN)、LSTM(long short-term memory)和BERT等基准模型,基于BERT-BiLSTM的短文本自动评分模型所获的二次加权kappa系数平均值最优. Aiming at the problems of sparse features,polysemy of one word and less context related information in short text automatic scoring,a short text automatic scoring model based on bidirectional encoder representations from transformers-bidirectional long short-term memory(BERT-BiLSTM)is proposed.Firstly,the large-scale corpus is pre-trained with bidirectional encoder representations from transformers(BERT)language model to acquire the semantic features of the general language.Then the semantic features of short text and the semantics of keywords in a specific context are acquired through the short text data for the pre-fine tuning downstream specific tasks set prefined by BERT.And then the deep-seated context dependency is captured through bidirectional long short-term memory(BiLSTM).Finally,the obtained feature vectors are input into Softmax regression model for automatic scoring.The experimental results show that compared with other benchmark models of convolutional neural networks(CNN),character-level CNN(CharCNN),long short-term memory(LSTM)and BERT,the short text automatic scoring model based on BERT-BiLSTM achieves the best average value of quadratic weighted kappa coefficient.
作者 夏林中 叶剑锋 罗德安 管明祥 刘俊 曹雪梅 XIA Linzhong;YE Jianfeng;LUO De’an;GUAN Mingxiang;LIU Jun;CAO Xuemei(Engineering Applications of Artificial Intelligence Technology Laboratory,Shenzhen Institute of Information Technology,Shenzhen 518172,Guangdong Province,P.R.China)
出处 《深圳大学学报(理工版)》 CAS CSCD 北大核心 2022年第3期349-354,共6页 Journal of Shenzhen University(Science and Engineering)
基金 广东省教育厅高校科研平台资助项目(2020KTSCX301) 深圳市基础研究计划资助项目(JCYJ20190808093001772) 国家高层次人才特殊支持计划领军人才(教学名师)资助项目(组厅字[2018]6号)。
关键词 信号与信息处理 自然语言处理 BERT语言模型 短文本自动评分 长短时记忆网络 二次加权kappa系数 signal and information processing natural language processing BERT language model short text automatic scoring long short-term memory net quadratic weighted kappa coefficient
  • 相关文献

参考文献4

二级参考文献21

共引文献110

同被引文献53

引证文献7

二级引证文献4

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部