期刊文献+

融合RoBERTa的多尺度语义协同专利文本分类模型 被引量:2

A multi-scale semantic collaborative patent text classification model based on RoBERTa
下载PDF
导出
摘要 针对静态词向量工具(如word2vec)舍弃词的上下文语境信息,以及现有专利文本分类模型特征抽取能力不足等问题,提出了一种融合RoBERTa的多尺度语义协同(RoBERTa-MCNN-BiSRU++-AT)专利文本分类模型。RoBERTa能够学习到当前词符合上下文的动态语义表示,解决静态词向量无法表示多义词的问题。多尺度语义协同模型利用卷积层捕获文本多尺度局部语义特征,再由双向内置注意力简单循环单元进行不同层次的上下文语义建模,将多尺度输出特征进行拼接,由注意力机制对分类结果贡献大的关键特征分配更高权重。在国家信息中心发布的专利文本数据集上进行验证,与ALBERT-BiGRU和BiLSTM-ATT-CNN相比,RoBERTa-MCNN-BiSRU++-AT部级专利的分类准确率分别提升了2.7%和5.1%,大类级专利的分类准确率分别提升了6.7%和8.4%。结果表明,RoBERTa-MCNN-BiSRU++-AT能有效提升对不同层级专利的分类准确率。 For patent text classification,the existing static word vector tools such as word2vec cannot express the context information of words,and most of the models can not completely extract features.Aiming at this problem,a multi-scale semantic collaborative patent text classification model based on RoBERTa,named RoBERTa-MCNN-BiSRU++-AT,is proposed.RoBERTa can learn the context-appropriate dynamic semantic representation of the current word and solve the problem that static word vectors cannot represent polysemous words.The multi-scale semantic collaboration model uses the convolution layer to capture the multi-scale local semantic features of text,and then uses the bidirectional built-in simple attention loop unit to model the context semantics at different levels.The multi-scale output features are spliced,and the key features that contribute more to the classification result are assigned higher weight by the attention mechanism.Experiments were carried out on the patent text data set published by the National Information Center.The results show that,compared with ALBERT-BiGRU and BiLSTM-ATT-CNN,RoBERTa-MCNN-BiSRU++-AT increases the accuracy by 2.7%and 5.1%respectively in patent text classification at the department level,and by 6.7%and 8.4%respectively in patent text classification at the major class level.RoBERTa-MCNN-BiSRU++-AT can effectively improve the classification effect of different levels of patent texts.
作者 梅侠峰 吴晓鸰 黄泽民 凌捷 MEI Xia-feng;WU Xiao-ling;HUANG Ze-min;LING Jie(School of Computer Science and Technology,Guangdong University of Technology,Guangzhou 510006,China)
出处 《计算机工程与科学》 CSCD 北大核心 2023年第5期903-910,共8页 Computer Engineering & Science
基金 广东省重点领域研发计划(2019B010139002) 广州市重点领域研发计划(202007010004)。
关键词 专利文本分类 语义协同 简单循环单元 RoBERTa模型 patent text classification semantic collaboration simple recurrent unit RoBERTa model
  • 相关文献

参考文献6

二级参考文献41

共引文献97

同被引文献21

引证文献2

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部