期刊文献+

CGGA:一种CNN与并行门控机制混合的文本分类模型 被引量:4

CGGA:Text Classification Model Based on CNN and Parallel Gating Mechanism
下载PDF
导出
摘要 针对中文文本分类准确率低、单一的卷积神经网络模型难以提取多方面特征的问题,本文提出一种基于CNN的并行门控机制的混合文本分类模型——CGGA(Convolutional Neural Network with parallel gating unit and attention mechanism).利用卷积提取文本的局部特征,并加入双向门控循环单元对数据进行上下文数据建模,提取关系特征,同时,引入门控Tanh-ReLU单元进行进一步的特征筛选,从而控制信息向下层流动的力度,并且减轻梯度弥散,提高模型分类准确率.最后,使用多头注意力机制进行权重更新计算,以提高在相应文本类别上的输出,进而优化模型分类性能.实验结果显示,本文提出的文本分类模型和分类算法,在THUCNews数据集和搜狐数据集上,比基线模型的宏平均精确率分别提高了2.24%、6.78%. For Chinese text classification accuracy rate is low,the single Convolutional Neural Networks model is difficult to extract multiple features,this paper proposes a hybrid text classification model based on parallel gating mechanism of CNN,named CGGA(Convolutional Neural Network with parallel gating unit and attention mechanism).Convolution is used to extract local features of the text,and the Bidirectional Gated Recurrent Unit is added to conduct context data modeling to extract relational features,at the same time,the Gated Tanh-ReLU Units was introduced for further feature screening, so as to control the dynamics of information flow to the lower layer,reduce the gradient dispersion and improve the classification accuracy of the model.Finally,the Multi-Head Aattention mechanism is used to calculate the weight update to improve the output of the corresponding text categories,so as to optimize the classification performance of the model.Experimental results show that the text classification model and classification algorithm proposed in this paper have improved the macro average precision of the THUCNews dataset and sohu dataset by 2.24% and 6.78%, respectively,compared with the baseline model.
作者 马建红 刘亚培 刘言东 陶永才 石磊 卫琳 MA Jian-hong;LIU Ya-pei;LIU Yan-dong;TAO Yong-cai;SHI Lei;WEI Lin(School of Software,Zhengzhou University,Zhengzhou 450002,China;School of Information Engineering,Zhengzhou University,Zhengzhou 450001,China;E-Government Center of Land and Resources of Henan Province,Zhengzhou 450002,China)
出处 《小型微型计算机系统》 CSCD 北大核心 2021年第3期516-521,共6页 Journal of Chinese Computer Systems
基金 科技部重点研发计划项目(2018YFB1701400)资助 郑州大学青年骨干教师培养计划项目(2017ZDGGJS048)资助。
关键词 卷积神经网络 门控Tanh-ReLU单元 双向门控循环单元 多头注意力机制 文本分类 Convolutional Neural Networks Gated Tanh-ReLU Units Bidirectional Gated Recurrent Unit Multi-Head Attention text classification
  • 相关文献

参考文献7

二级参考文献73

共引文献442

同被引文献31

引证文献4

二级引证文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部