期刊文献+

一种轻量级文本蕴含模型 被引量:3

A lightweight text entailment model
下载PDF
导出
摘要 现有主流文本蕴含模型大多采用循环神经网络编码,并采用各种注意力推理机制或辅以手工提取的特征来提升蕴含关系识别准确率,由于复杂的网络结构和RNNs网络串行机制导致这些模型训练和推理速度较慢.本文提出轻量级文本蕴含模型,采用自注意力编码器编码文本向量,点积注意力交互两段文本,再采用卷积神经网络对交互特征推理,整个结构可根据不同数据的推理难度叠加不同模块数量.在多个文本蕴含数据集实验表明,本文模型在保持较高识别准确率情况下仅用一个块参数仅为665 K,模型推理速度相比其他主流文本蕴含模型至少提升一倍. Most of the existing mainstream textual entailment models adopt recurrent neutral network to encode text,and various complex attention mechanisms or manually extracted text features are used to improve the accuracy of textual entailment recognition.The training and inference speed of the models is usually slow due to the complex network structure and the sequential nature of RNNs.In this paper,Lightweight Text Entailment Model is proposed.In the proposed model,the self-attentional encoder is adopted to encode text vectors;the dot product attention mechanism is adopted to interact two texts;the convolutional neural network is adopted to deduce interactive features,and the module number of the structure can be adjusted according to the reasoning difficulty of data.Experiments on multiple datasets show that the parameter size of single module in the model is only 665 K,and the inference speed of the model is at least twice as high as that of other mainstream models,under the condition of high accuracy.
作者 王伟 孙成胜 伍少梅 张芮 康睿 李小俊 WANG Wei;SUN Cheng-Sheng;WU Shao-Mei;ZHANG Rui;KANG Rui;LI Xiao-Jun(China Electronic Technology Cyber Security Company Limited,Chengdu 610041,China;College of Computer Science,Sichuan University,Chengdu 610065,China;Westone Information Industry INC,Chengdu 610041,China)
出处 《四川大学学报(自然科学版)》 CAS CSCD 北大核心 2021年第5期31-38,共8页 Journal of Sichuan University(Natural Science Edition)
基金 四川省新一代人工智能重大专项(2018GZDZX0039) 四川省重点研发项目(2019YFG0521,JG2020125)。
关键词 注意力机制 卷积神经网络 轻量级 文本蕴含 Attention mechanism CNN Lightweight Textual entailment
  • 相关文献

参考文献3

二级参考文献8

共引文献47

同被引文献24

引证文献3

二级引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部